Next Article in Journal
Recognition and Counting of Apples in a Dynamic State Using a 3D Camera and Deep Learning Algorithms for Robotic Harvesting Systems
Next Article in Special Issue
Dynamic Resource Optimization for Energy-Efficient 6G-IoT Ecosystems
Previous Article in Journal
Rapid and On-Scene Chemical Identification of Intact Explosives with Portable Near-Infrared Spectroscopy and Multivariate Data Analysis
Previous Article in Special Issue
Customer Analysis Using Machine Learning-Based Classification Algorithms for Effective Segmentation Using Recency, Frequency, Monetary, and Time
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Video Process Mining and Model Matching for Intelligent Development: Conformance Checking

School of Computer Science and Engineering, Shandong University of Science and Technology, Qingdao 266000, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(8), 3812; https://doi.org/10.3390/s23083812
Submission received: 13 March 2023 / Revised: 28 March 2023 / Accepted: 29 March 2023 / Published: 7 April 2023
(This article belongs to the Special Issue Artificial Intelligence and Advances in Smart IoT)

Abstract

:
Traditional business process-extraction models mainly rely on structured data such as logs, which are difficult to apply to unstructured data such as images and videos, making it impossible to perform process extractions in many data scenarios. Moreover, the generated process model lacks analysis consistency of the process model, resulting in a single understanding of the process model. To solve these two problems, a method of extracting process models from videos and analyzing the consistency of process models is proposed. Video data are widely used to capture the actual performance of business operations and are key sources of business data. Video data preprocessing, action placement and recognition, predetermined models, and conformance verification are all included in a method for extracting a process model from videos and analyzing the consistency between the process model and the predefined model. Finally, the similarity was calculated using graph edit distances and adjacency relationships (GED_NAR). The experimental results showed that the process model mined from the video was better in line with how the business was actually carried out than the process model derived from the noisy process logs.

1. Introduction

Business processes store information about the execution of information systems in event logs, and process mining extracts useful information from the event logs [1]. Process mining’s purpose is to compare the event logs or the resulting process model in the discovery task with the existing reference model of the same process and to detect issues in the executed process for improvement [2]. The six degrees of parallel platform is a widely used parallel mechanism that can achieve six DOF of motion of a moving platform in space [3]. This study aims to enhance the security of people’s health, improve the medical level further, and increase the confidentiality of people’s private information [4]. Traditional process mining techniques are based on process logs. However, logs cannot be obtained effectively in some scenarios so it is impossible to mine process models that conform to actual scenarios. In addition, some process logs may contain some erroneous or incomplete behaviors [5], which cause a certain deviation between the model obtained by the process mining algorithm and the ideal model. A method of extracting the process models from videos and analyzing the consistency of the process models is proposed.
Due to various problems in the process logs and a large amount of video data emerging [6], it is very necessary to apply process mining to video data. In the era of big data, the Internet is enmeshed in people’s lives and brings conveniences to their production and lives [7]. Most enterprises have complete video data, which can completely record the actual execution of the business process and contains rich business information. The conformance checking of procedural video is based on video data as the source data, and the process model excavated and the process model obtained from the process log containing noise are checked for consistency with the predefined models so that enterprises can improve their production scheduling and achieve a high-level management of business processes [8]. In addition, it can optimize resources, reduce the operating costs of enterprises, and make enterprises obtain greater profits. Compared with process mining using the process logs as data sources, the work of extracting the processes from videos is less. One of the main reasons is the complexity of video processing technology. The continuous development of moving target detection methods such as the self-supervised video [9], the trapezoid normalized pixel [10], the cycleGan [11], and multi-agent systems [12] provide technical support for video moving target detection. A video is a combination of continuous images. Mining potential task sequences in videos face various problems, for example, most videos suffer from frequent changes in lighting, occlusion between objects, etc. The existing video anomaly detection methods can only detect the anomaly of a single behavioral event but do not consider whether the actual execution of the business process in the video conforms to the actual situation. Traditional methods usually rely on good initialization and find difficulty in handling large interframe transformations due to fast camera motion [13]. Yet, there is no framework in place to evaluate the current state of the deployment of digital education at the beginning of the educational process, orient public education policies, and establish a baseline for comparison and future development among developing countries, ensuring that their future workforce is ready for the challenges of the twenty-first century. To this end, a procedural video conformance-checking method based on process mining and model matching is proposed.
The organization of this paper is as follows: Section 2 introduces the related work on process mining; Section 3 gives the framework of the paper and the introduction of key algorithms for each functional module; Section 4 is the specific experiments and analysis; and Section 5 summarizes the work of this paper.

2. Related Work

In recent years, process mining has gained extensive attention and accumulated more and more research results. A process mining anonymization algorithm based on genetic algorithms was proposed in [14], which reduces the privacy protection and utility tradeoff problem of process mining to the optimization problem of the activity inhibition set. Synthesizing photo-realistic images from text descriptions is a challenging problem in computer vision and has many practical applications [15]. In [16] they gave a technique to repair the missing events in a log. They used the knowledge gathered from the process model and provided a technique to repair the missing event logs in a trace. This technique gives us an incomplete log analysis. By using the stochastic Petri net model and alignment, we repair the event logs and convert them into Bayesian analysis. An efficient algorithm for mining the nearest neighbor relationships of relational event logs was proposed in [17]. The algorithm used two-level pointers to scan the log data, which has a linear time complexity for the log size, and the demand for memory space was independent of the log size. This algorithm improved the process mining efficiency of relational event logs.
A model mining method of cross-organizational emergency response process (CERP) was proposed in [18]. This method first extended the classical Petri nets (RMPNs) with resource and message attributes, and found the emergency response process models (IERP) within the organization were represented as RMPNs from the emergency exercise process logs. Secondly, it defined the cooperation mode between emergency organizations. Finally, the CERP model was obtained by combining the IERP model with the collaboration model. Text mining techniques were employed to mine the sentiments in different interactions, and then epistemic network analysis (ENA) was used to uncover sentiment changes in the five learning stages of blended learning [19]. In [20], they formalized the problem of the directly-follows graph (DFG) simplification as an optimization, specifically, the problem of finding a sound-spanning subgraph of a DFG with a minimal number of edges and a maximal sum of edge frequencies. Finally, they reported on an evaluation of the efficiency and optimality of the proposed heuristics using 13 real-life event logs.
To sum up, these traditional process discovery methods used process logs as input. However, process logs cannot be obtained effectively in some scenes, which lead to the process extraction work facing huge challenges. Relatively speaking, unstructured video data are more flexible, records the real situation of the business process, and objectively reflects the actual execution process of the process.

3. Framework

Based on the above analysis, a method for extracting process models from videos and performing conformance checking is proposed, which solves the problems where the process logs cannot be obtained effectively and the process logs are noisy. The framework of our work is shown in Figure 1, including the video action record extraction module and the video conformance checking module. The video action record extraction module first preprocesses the video data, removes image noise and extracts moving targets, and then inputs the preprocessed images into the action localization network to generate action suggestions of different length intervals, and finally identifies the action category label of the action suggestion interval through the action recognition network, and saves the action category label in the log, builds a video flow log, adds some noise on this basis, such as wrong or incomplete behavior, and builds a video flow log with the noise; the purpose of adding noise is to simulate actual process logs with noise in the information system. The video conformance checking module includes generating predefined models and conformance checking. The predefined module defines the standard Petri net process models, and the conformance checking module uses the graph edit distance and the adjacency relationship (GED_NAR) method to judge the matching degree of the process model containing noise and the extracted process model with the predefined model.

3.1. Video Action Record Extraction Module

The action record extraction of the video includes two parts: video data preprocessing, and action localization and recognition. The goal is to extract process information from the process video. Video data preprocessing can maximize data simplification and reduce computational interference. Motion localization and recognition is the process of filtering and identifying preprocessed images from the video data to complete the process extraction of procedural video.

3.1.1. Video Data Preprocessing

Four techniques are used for the video preprocessing: image grayscale processing, knn-based moving target extraction, image noise processing, and open operation.
(1)
Image grayscale processing: The formula is shown in (1), where H and W represent the height of the original video frame and width, i ∈ [1…H], j ∈ [1…W].
G r a y = m a x { R ( i , j ) , G ( i , j ) , B ( i , j ) }
Image grayscale can turn three channels into a single channel, reduce memory space, improve operation speed, increase visual contrast, and highlight the target area.
(2)
Moving target extraction: The algorithm’s purpose is to extract the motion information contained in the action in the video. The image is grayed out and still contains useless background information. In order to accurately extract the moving targets, the moving target is extracted by the knn algorithm [21]. Basically, the knn algorithm can accurately extract moving targets, but it is still disturbed by noise.
(3)
Image noise processing: The method of median filtering is used to perform nonlinear filtering on grayscale images, so that the target pixel is closer to the real value, thereby eliminating isolated noise points. The calculation method is shown in Formula (2), where g (x, y) is the processed image, f (x, y) is the original image, W is an N*N two-dimensional template, N is usually a positive odd number, and Med represents sorting the gray values in the domain window and taking out the middle value:
g ( x , y ) = M e d { f ( x k , y l ) , ( k , l W ) }
(4)
Open operation processing: Assuming that Z is the target image and W is a structural element, the mathematical formula for the opening operation processing of the structural element W by the target image Z is:
Z W = ( Z Θ W ) W
Z Θ W = x , y W x y Y
Z W = { x , y | ( W ) x y X }
where (W)xy represents the translation of the origin of the structuring element W to the position of the image pixel (x, y), the erosion operation is represented by Θ, and the dilation operation is represented by ⨁. The open operations smooth the contours of the moving targets and break up the narrow connection areas and remove the small protrusions from targets.
The video data preprocessing can extract the interest points of the actual actions and high-light the potential task sequences in the video. The processing results are shown in Figure 2.

3.1.2. Action Location and Recognition

After video data preprocessing, the next step is to locate and recognize actions in the preprocessed image sequence, as shown in Figure 3. For videos, both the target action and the duration of the action are varied. The coherence between actions makes it difficult to locate the start and end points of the action. A binary classification method based on a convolutional neural network is used to distinguish between action and background. To generate the temporal region proposals, our basic idea is to group consecutive snippets with high actionness scores. In a given video, we first sample a sequence of snippets, then use this network to produce actionness scores for them, and finally group them into temporal regions of various granularities. A fault-tolerant processing scheme with high robustness is designed, which can generate action suggestion intervals of different lengths, allowing accurate identification of the duration of the action in the video. With a set of proposed temporal regions, the next stage is to classify them into action classes.
In short, we hope that the action proposed interval can cover all kinds of action time. A binary classification network is designed to distinguish action from the background, filtering out which clips contain action and which clips are backgrounds. Firstly, input the preprocessed image into the binary classification network, and each image gets a binary classification result of 0 or 1, where 0 means the image is the background, and 1 means the image is of an action. Then, a fault-tolerant mechanism is established to segment action suggestion intervals of different lengths, and the confidence of these images is taken as the score of the action suggestion interval. Finally, the intersection over the union of the action suggestion interval and the basic fact are calculated, and the non-maximum suppression algorithm is adopted. The location of the chopping board action instance is shown in Figure 4. The green box represents ground truths, the blue box represents a good prediction, and the red box represents a poor prediction.
A robust fault-tolerant mechanism is designed by us, which is controlled by two design parameters: the score threshold τ and the tolerance threshold γ. If the confidence of the image is greater than τ, the image is marked as ‘1’, indicating that the image contains action. If the confidence of the image is less than τ, the image is marked as ‘0’, indicating this image is the background. For the fault tolerance threshold γ, we choose an image as a starting point, and recursively expand it by absorbing subsequent images, and terminate the expansion if the proportion of action suggestion interval label ‘1’ is less than γ. Obviously, this multi-threshold design enables us to remove background clips, generate action propose intervals of different lengths, and improve the accuracy of the video action location. As shown in Figure 5, the purple arrow indicates the extension direction of the action interval range.
Compared with images, a video not only needs to consider the spatial location information of the video frames but also needs to consider the temporal relationship between the video frames. Pretrained models on ImageNet [22] have achieved great success in the field of analog images, extending the 2D convolution in the picture to the 3D convolution in the video, and adding the idea of two streams to the 3D convolution network. [23] proposed two-stream inflated 3D convolutional neural network (Two-Stream I3D), where the spatial stream network extracts features of the objects from static RGB video frames and the temporal stream network extracts the motion information; finally, the outputs of these two layers are fused to obtain the action recognition results. A new dynamic sign language recognition network integrating dual stream 3D convolutional neural networks and attention mechanisms is proposed by [24]. The convolutional block attention module (CBAM) [25] is introduced into the I3D network to enable the network to learn the salient information in the image and make the important features in the image more prominent without affecting the performance of the original network. However, considering that the spatial location information and time-domain relationships are two different concepts, the learning methods of the two levels must be different for neural networks. Using the same network structure for spatial and temporal streams will lead to more redundant information in the stream information of the two layers in the final fusion stage, which wastes a large number of neuron parameters of the dual-stream convolutional network.
Based on the above analysis, a network model integrating two-stream heterogeneous 3D convolution and attention mechanism is proposed, which introduces CBAM. The input is the original RGB image and the preprocessed optical flow map. The entire network model is shown in Figure 6. The spatial channel is added with CBAM after the Concatenation layer of the Inception module of I3D; the network connection method is shown in Figure 6a. The time channel only uses the Inception module of the I3D network, and also adds CBAM after the Concatenation layer. The network connection method is shown in Figure 6b. In addition to adding the attention mechanism CBAM, the spatial channel also improves the I3D network structure by: (1) Removing the first max pooling layer to prevent the loss of low-level features of the image; (2) Removing the final average pooling operation, so that the global information in the image can be preserved. Using the dual-stream heterogeneous network not only comprehensively considers the rich spatial feature information and motion features in the video frame, but also reduces the neural network parameters and improves the computational efficiency of the model.
The loss function used for network training is the cross-entropy function.
l o s s = i n L ( i ) l o g S ( i )
where n represents the total action label category, L represents the OneHot encoding of the true label, and S represents the probability of the predicted action label output by the Softmax layer. In the training process, the Backpropagation Algorithm is used to continuously update the network parameters to reduce the loss, and finally, the argmax function is used to obtain the category label of the action according to the maximum probability value.

3.2. Video Consistency Analysis Module

Video conformance checking includes predefined model and conformance checking. Predefined modules define standard Petri net process models; consistency is a calculation of the fit of process models extracted from noisy process logs and videos from predefined models.

3.2.1. Predefined Model

After analysis of the dataset, each process executor performs a series of actions in a natural state, and then a Petri net is used to represent a predefined model of the process, as shown in Figure 7. The specific descriptions of the events in Figure 7 are shown in Table 1. The following is a description of the predefined model where the process executor enters the scene and then has two options:
(1)
Take the cutting board and rag, then put the cutting board and rag in order, then take the plate and cup, then put the plate and cup in order, then there are two options: ➀ Take a fork, knife, spoon, and put the fork, knife, spoon in that order; ➁ Take a fork, a spoon, and a knife, and then put the fork, spoon, and knife in that order.
(2)
Take the cutting board, then put the cutting board, then take the rag, then put the rag, then take the plate, then put the plate, then take the spoon, then put the spoon, then open the cupboard, take the cup, put the cup, or just take the cup, put the cup.
No matter which utensils the executor sets first, the last process executor will leave the scene.

3.2.2. Conformance Checking

In order to verify that the process model extracted from the video is more consistent with the actual execution of the business than the process model obtained from the process log containing noise, we did a comparative experiment to calculate the fitness of the two process models and the predefined model. Firstly, convert the process model extracted from the video, the process model obtained from the process log containing noise, and the predefined model into a directed graph, and then use the GED_NAR algorithm to calculate the fitness of the directed graph. Finally, the compliance results of the two process models and the predefined model are obtained.
With action recognition, we can get the labels of each action category in the video, and after setting a unique ID for each video, we can build a CSV file containing two columns of attributes: “video ID” and “action label”. The “video ID” can represent the “case ID” in the process log, and the “action label” can represent the “activity” content in the process log. Finally, the video data process log file is obtained, and ProM is used to generate the Petri net model, as shown in Figure 8.
The relevant definitions of the GED_NAR method are given below:
Definition 1.
The fitness calculation of two directed graphs G1 = (N1, E1) and G2 = (N2, E2) is shown in Formula (7) where wskipn represents the cost coefficient of inserting and deleting action nodes, and wskipe represents the cost coefficient of inserting and deleting action relation edges; the value range is [0,1], which can be set according to actual needs with the default value set to 1.
F G E D ( G 1 , G 2 ) = 1.0 ( w s k i p n × f s k i p n + w s k i p e × f s k i p e ) / ( w s k i p n + w s k i p e )
where fskipn represents the cost function of inserting and deleting action nodes, as shown in Formula (8), skipn represents the set of inserting and deleting action nodes, N1 and N2 represent the number of action nodes of the model.
f s k i p n = ( | s k i p n | ) / ( | N 1 | + | N 2 | )
where fskipe represents the cost function of inserting and deleting action relation edges, as shown in Formula (9), skipe represents the edge set of inserting and deleting, E1 and E2 represent the number of action relation edges of the model.
f s k i p e = ( | s k i p e | ) / ( | E 1 | + | E 2 | )
Definition 2.
If G = (N,E) is a directed graph, there are two nodes p, with qN satisfying e = <p, q >E, then < p, q> is called a node adjacency relationship, abbreviated as NAR, for a given directed graph. All node adjacencies constitute an adjacency set, denoted as NARs = {<p, q > |e = <p, q> ∈ E}.
The degree of fitness based on the adjacency relationship is calculated as shown in Formula (10):
F N A R ( G 1 , G 2 ) = ( N A R s 1 N A R s 2 ) / ( N A R s 1 N A R s 2 )
Definition 3.
If M1 is a predefined model, M2 is an extracted process model, and G1 and G2 are directed graphs of M1 and M2, then the formula for the degree of fitness between the process model and the predefined model is shown in (11):
F i t n e s s G E D _ N A R ( M 1 , M 2 ) = α × F G E D G 1 , G 2 + σ × F N A R ( G 1 , G 2 )
where α and σ are two coefficients, and the value range is [0,1], which can be set according to actual needs with the default value of 0.5 and α + σ = 1.

4. Experiments

Experiments analyzed the dataset, action localization, action recognition, and conformance checking.

4.1. Dataset

At present, most of the datasets only involve action localization and action recognition, and there are very few datasets containing process information, so we use the public dataset TUM [26] for experiments, as shown in Figure 9. In the video, after the process executor enters the monitoring screen, he starts to take the cutting board and put it on the table. After the tableware is placed in an orderly manner, he finally leaves the monitoring scene. The duration of each video in the data set is in intervals of 1–2 min, and the action types in the video can be divided into taking a plate, placing a plate, etc. The “background class” is without any action. Finally, the video frame sequence is used to train and test the action localization and recognition networks.

4.2. Experiment of Action Localization

The action localization method is to design a two-class network to filter the video frames containing actions. The indicators used are Iou, Recall, Precision, and Aver-age-Precision. The final experimental results are shown in Table 2. We can find that the action localization scheme designed in this paper can locate the starting position and ending position of the action very well, which is mainly due to the fact that we not only set an action threshold to indicate how many points are counted as actions but also set a tolerance threshold to prevent the interference of noise, that is, the appearance of several video frames with background in consecutive frames still adds to the clip. Therefore, the action localization model designed in this paper has better results.

4.3. Experiment of Action Recognition

We compare our model with other methods in Table 3, including LRCN [27], 3D-ConvNet [28], Two-StreamI3D [23], Two-Stream-CBAM-I3D [24], and AFSD [29]. The metrics used are accuracy, precision, recall, and a weighted composite evaluation index of the three for the experimental evaluation of action recognition. To avoid interference of other factors, the hyperparameter settings such as the learning rate, training rounds, and times used by the models are kept uniform. The results are shown in Table 3 and Figure 10. The various indicators of the fusion dual-stream heterogeneous 3D convolution and attention mechanism network model proposed in this paper are significantly improved compared to other models. Firstly, compared with the original Two-Stream I3D model, the accuracy rateincreased by 5%, the precision rate increased by 13.4%, the recall rate increased by 14.1%, and the average rate increased by 10.8%. Secondly, the two-stream heterogeneous structure reduces the complexity of the model and shortens the training time. Compared with the dual-stream CBAM-I3D model, the indicators are also slightly improved. Finally, the optical flow graphs generated by the preprocessing of the video data for the network inputs are improved to varying degrees. The analysis of the reasons for this shows that the network model proposed not only retains the originality of the input video information but also improves the expressive ability of the network, and automatically learns the importance of the spatial position information and temporal relationship of video frames, then according to the degree of importance, key features are enhanced and useless features are suppressed, so as to highlight the saliency information in the video frame. The introduction of the CBAM attention mechanism enables the network to learn more important spatial and temporal features in video frames without affecting the original network. The two-stream heterogeneous mode avoids repeated extraction of information, reduces the number of model calculation parameters and model size, and makes network convergence faster.

4.4. Conformance Checking Experimental Analysis

Conformance checking analyzes how the process model matches the predefined model. In addition, to convert the process model into a directed graph with reference to the conversion rules proposed in [30] (as shown in Figure 11), the GED_NAR method is used to calculate the fitness of the model.
According to the Definition 1 of the GED_NAR method, the conversion of G to G1 requires deleting two nodes (t4, t5), deleting (<t3, t4> <t4, t5> <t5, t6>) three edges and adding (<t3, t6>) a side. Then the calculation process is as follows: |skipn| = 2, |skipe| = 4, fskipn = 2/(21 + 19) = 0.05, fskipe = 4/(23 + 21) = 0.09, FGED = 1.0 − (1 × 0.05 + 1 × 0.09)/(1 + 1) = 0.93.
Converting G to G2 requires deleting four nodes (t4, t5, t20, t21), deleting < t3, t4> <t4, t5> <t5, t6> <t9, t10> <t16, t20> <t20, t21> <t21, t12>) seven edges and adding (<t3, t6>) one edge. Then the calculation process is as follows: |skipn| = 4, |skipe| = 8, fskipn = 4/(21 + 17) = 0.11, fskipe = 8/(23 + 17) = 0.20, FGED = 1.0−(1 × 0.11 + 1 × 0.20)/(1 + 1) = 0.85.
According to the Definition 2 of the GED_NAR method, the node adjacency set of G is: NARs = {<t1, t2> <t2, t3> <t3, t4> <t4, t5> <t5,t6>…… <t16, t20> <t20, t21> <t21, t12>}. The node adjacency set of G is: NARs1 = {<t1, t2> <t2, t3> <t3, t6>……<t16, t20> <t20, t21> <t21, t12>}. The corresponding intersection relation is: NARsꓵNARs1 ={<t1, t2> <t2, t3>……<t16, t20> <t20,t21> <t21, t12>}; NARsUNARs1 = {<t1, t2> <t2, t3> <t3,t4> <t3, t6> <t4, t5>……<t16, t20> <t20,t21> <t21, t12>}. Therefore, the fitness degree of G and G1 obtained based on the node adjacency set is FNAR(G, G1) = 20/24 = 0.83.
According to the Definition 2 of the GED_NAR method, the node adjacency set of G2 is: NARs2 = {<t1, t2> <t2, t3> <t3, t6>……<t16, t17> <t17, t18> <t18, t12>}; The corresponding intersection relation is: NARsꓵNARs2 = {<t1, t2> <t2, t3>……<t16, t17> <t17, t18> <t18, t12>}; NARsꓴNARs2 = {<t1, t2> <t2, t3> <t3, t4> <t3, t6> <t4, t5> <t5, t6>……<t16, t20> <t20, t21> <t21, t12>}. Therefore, the fitness degree of G and G2 obtained based on the node adjacency set is: FNAR(G, G2) = 16/24 = 0.67.
According to the Definition 3 of the GED_NAR method, the process model extracted from the video and the predefined model can be obtained by using GED_NAR as:
FitnessGED-NAR(G, G1) = 0.5 × FGED(G, G1) + 0.5 × FNAR(G, G1) = 0.5 × 0.93 + 0.5 × 0.83 = 0.88.
According to the Definition 3 of the GED_NAR method, the process model obtained from the noisy process log and the predefined model can be obtained by using GED_NAR as:
FitnessGED-NAR(G, G2) = 0.5 × FGED(G, G2) + 0.5 × FNAR(G, G2) = 0.5 × 0.85 + 0.5 × 0.67 = 0.76.
From the experimental results, it can be seen that the fitness of the process model and the predefined model extracted from the video is greater than the fitness of the process model and the predefined model obtained from the process log containing noise. The experiment results show that the process model extracted from the video is more consistent with the actual execution process of the business.

5. Conclusions

Aiming at the problem that the process logs cannot be obtained effectively and the process logs generated by the information system will have noise, a method for extracting process models from videos and analyzing their consistency is proposed by us. Firstly, the video data preprocessing removes the background information irrelevant to the moving target in the video picture and only retains the spatiotemporal interest point area. Secondly, a binary classification network is used to generate action suggestion intervals of different lengths, which are then inputed to the fusion dual-stream heterogeneous, the three-dimensional convolution and attention mechanism network to classify actions. Finally, consistency detection is carried out with the predefined model and with the process model obtained from the process logs containing noise and the process model extracted from the video.
The method proposed in this paper is only for single-person video process modeling, and cannot handle multi-person collaborative business processes well. The next step is to study how to model a multi-person collaborative video process and conduct conformance checking.

Author Contributions

Writing—original draft preparation, S.C.; writing—review and editing, M.Z.; R.C.; Z.Z. and Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by NSFC (U1931207 and 61702306), Sci. & Tech. Development Fund of Shandong Province of China (ZR2022MF288, ZR2017BF015, and ZR2017MF027), the Taishan Scholar Program of Shandong Province, and SDUST Research Fund (2015TDJH102 and 2019KJN024).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Enquiries about data availability should be directed to the authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, C.; Cheng, L. Process mining: An overview of technologies and applications. China Sci. Technol. Ind. 2021, 9, 50–51. [Google Scholar] [CrossRef]
  2. Heidy, M.; Edgar, T. Event Log Preprocessing for Process Mining: A Review. Appl. Sci. 2021, 11, 10556. [Google Scholar] [CrossRef]
  3. Liu, M.; Gu, Q.; Yang, B.; Yin, Z.; Liu, S.; Yin, L.; Zheng, W. Kinematics Model Optimization Algorithm for Six Degrees of Freedom Parallel Platform. Appl. Sci. 2023, 13, 3082. [Google Scholar] [CrossRef]
  4. Lv, Z.; Yu, Z.; Xie, S.; Alamri, A. Deep Learning-Based Smart Predictive Evaluation for Interactive Multimedia-Enabled Smart Healthcare. ACM Trans. Multimedia Comput. Commun. 2022, 18, 1–20. [Google Scholar] [CrossRef]
  5. Zeng, Q.T. Summary of research status and problems in process mining. J. Syst. Simul. 2007, S1, 275–280. [Google Scholar]
  6. Chen, J.Q. The development trend of china’s video sharing market in the 5G era. Youth Report. 2020, 12, 88–89. [Google Scholar] [CrossRef]
  7. Qin, X.; Liu, Z.; Liu, Y.; Liu, S.; Yang, B.; Yin, L.; Liu, M.; Zheng, W. User OCEAN Personality Model Construction Method Using a BP Neural Network. Electronics 2022, 11, 3022. [Google Scholar] [CrossRef]
  8. Huang, L.H. Research and Application of Business Law Information Management System Based on Big Data. J. Sens. 2022, 2022, 1–12. [Google Scholar] [CrossRef]
  9. Liu, Y.; Wang, K.; Liu, L.; Lan, H.; Lin, L. Tcgl: Temporal contrastive graph for self-supervised video representation learning. IEEE Trans. Image Process. 2022, 31, 1978–1993. [Google Scholar] [CrossRef] [PubMed]
  10. Xu, S.; He, Q.; Tao, S.; Chen, H.; Chai, Y.; Zheng, W. Pig Face Recognition Based on Trapezoid Normalized Pixel Difference Feature and Trimmed Mean Attention Mechanism. IEEE Trans. Instrum. Meas. 2022, 1, 3500713. [Google Scholar] [CrossRef]
  11. Li, B.; Lu, Y.; Pang, W.; Xu, H. Image Colorization using CycleGAN with semantic and spatial rationality. Multimed. Tools Appl. 2023. [Google Scholar] [CrossRef]
  12. Duan, J.; Duan, G.; Cheng, S.; Cao, S.; Wang, G. Fixed-time time-varying output formation–containment control of heterogeneous general multi-agent systems. ISA Trans. 2023. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, J.; Tang, Y.; Wang, H.; Xu, K. ASRO-DIO: Active Subspace Random Optimization Based Depth Inertial Odometry. IEEE Trans. Robot. 2022, 39, 1496–1508. [Google Scholar] [CrossRef]
  14. Birgin, M.K.; Bilge, A. Genetic Algorithm-Based Privacy Preserving Collaborative Filtering. In Proceedings of the 2019 Innovations in Intelligent Systems and Applications Conference (ASYU), Izmir, Turkey, 31 October–2 November 2019. [Google Scholar] [CrossRef]
  15. Liu, H.; Xu, Y.; Chen, F. Sketch2Photo: Synthesizing photo-realistic images from sketches via global contexts. Eng. Appl. Artif. Intell. 2023, 117, 105608. [Google Scholar] [CrossRef]
  16. Shabnam, S.; Fang, X.W.; Usman, A. Repairing Event Logs to Enhance the Performance of a Process Mining Model. Math. Probl. Eng. 2022, 43, 515–527. [Google Scholar] [CrossRef]
  17. Gao, J.; Liu, C.; Liu, Y.F. An Efficient Method for Mining Immediate Neighborhood Relationships in Relational Databases for Event Logs Computer Integrated Manufacturing System. Comput. Integr. Manufac-Turing Syst. 2020, 26, 1492–1499. [Google Scholar]
  18. Liu, C.; Li, H.L.; Zeng, Q.T.; Li, C.H. Cross-Organization Emergency Response Process Mining: An Approach Based on Petri Nets. Math. Probl. Eng. 2020, 52, 8836007. [Google Scholar] [CrossRef]
  19. Huang, C.; Han, Z.; Li, M.; Wang, X.; Zhao, W. Sentiment evolution with interaction levels in blended learning environments: Using learning analytics and epistemic network analysis. Australas. J. Educ. Technol. 2021, 37, 81–95. [Google Scholar] [CrossRef]
  20. Hugo, H.D.; Augusto, V.; Jouaneton, B.; Lamarsalle, L. An optimization-based process mining approach for explainable classification of timed event logs. In Proceedings of the 2020 IEEE 16th International Conference on Automation Science and Engineering, Hong Kong, China, 20–21 August 2020; pp. 43–48. [Google Scholar] [CrossRef]
  21. Zivkovic, Z.; Heijden, F. Efficient adaptive density estimation per image pixel for the task of background subtraction. Pattern Recognit. Lett. 2006, 27, 773–780. [Google Scholar] [CrossRef]
  22. Deng, J.; Dong, R.; Socher, L.J. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; Volume 28, pp. 248–255. [Google Scholar] [CrossRef] [Green Version]
  23. Carreira, J.; Zisserman, A. Quo vadis, action recognition? A new model and the kinetics dataset. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, Hawaii, 21–26 July 2017; pp. 6299–6308. [Google Scholar] [CrossRef] [Green Version]
  24. Wang, F.H.; Zhang, Q.; Huang, C.; Zhang, L. Dynamic gesture recognition incorporating dual-stream 3D convolution and attention mechanism. Comput. Intell. Neurosci. 2021, 43, 1389–1396. [Google Scholar] [CrossRef]
  25. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  26. Tenorth, M.; Bandouch, J.; Beetz, M. The TUM kitchen data set of everyday manipulation activities for motion tracking and action recognition. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision Workshops, Kyoto, Japan, 27 September–4 October 2009; pp. 1089–1096. [Google Scholar] [CrossRef] [Green Version]
  27. Donahue, J.; Hendricks, L.A.; Guadarrama, S.; Rohrbach, M.; Venugopalan, S. Long-Term Recurrent Convolutional Networks for Visual Recognition and Description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, Hawaii, 21–26 July 2017; pp. 2625–2634. [Google Scholar] [CrossRef]
  28. Le, T.M.; Lnoue, N.; Shinoda, K. A Fine-to-Coarse Convolutional Neural Network for 3D Human Action Recognition. In Proceedings of the International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 1984–1987. [Google Scholar] [CrossRef]
  29. Lin, C.; Xu, C.; Luo, D. Learning salient boundary feature for anchor-free temporal action localization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3320–3329. [Google Scholar] [CrossRef]
  30. Remco, D.; Dumas, M.; GarciaBanuelos, L. Graph matching algorithms for business process model similarity search. In Proceedings of the International Conference on Business Process Management, Ulm, Germany, 8–10 September 2009; pp. 48–63. [Google Scholar]
Figure 1. The framework of our approach.
Figure 1. The framework of our approach.
Sensors 23 03812 g001
Figure 2. Sample Image of video data preprocessing (a) Video frame (b) Grayscale image processing image (c) Target extraction processing image (d) Median filter processing image (e) Open operation processing image.
Figure 2. Sample Image of video data preprocessing (a) Video frame (b) Grayscale image processing image (c) Target extraction processing image (d) Median filter processing image (e) Open operation processing image.
Sensors 23 03812 g002
Figure 3. Action location and recognition.
Figure 3. Action location and recognition.
Sensors 23 03812 g003
Figure 4. Action instance. The green box represents ground truths, the blue box represents a good prediction, and the red box represents a poor prediction.
Figure 4. Action instance. The green box represents ground truths, the blue box represents a good prediction, and the red box represents a poor prediction.
Sensors 23 03812 g004
Figure 5. Schematic image of fault tolerance.
Figure 5. Schematic image of fault tolerance.
Sensors 23 03812 g005
Figure 6. (a) CBAM-I3D network (b) CBAM-Inception network (c) Fusion two-stream heterogeneous 3D convolution and attention mechanism network.
Figure 6. (a) CBAM-I3D network (b) CBAM-Inception network (c) Fusion two-stream heterogeneous 3D convolution and attention mechanism network.
Sensors 23 03812 g006
Figure 7. Example image of the predefined model part.
Figure 7. Example image of the predefined model part.
Sensors 23 03812 g007
Figure 8. Flow chart of Petri net generation.
Figure 8. Flow chart of Petri net generation.
Sensors 23 03812 g008
Figure 9. Sample Dataset.
Figure 9. Sample Dataset.
Sensors 23 03812 g009
Figure 10. Performance comparison of action recognition.
Figure 10. Performance comparison of action recognition.
Sensors 23 03812 g010
Figure 11. G: The directed graph corresponding to the predefined model, G1: The directed graph corresponding to the extracted process model, G2: The directed graph corresponding to the process model with noise.
Figure 11. G: The directed graph corresponding to the predefined model, G1: The directed graph corresponding to the extracted process model, G2: The directed graph corresponding to the process model with noise.
Sensors 23 03812 g011
Table 1. Correspondence table of event number and name.
Table 1. Correspondence table of event number and name.
Event SequenceEvent NameEvent SequenceEvent Name
t1Enter the scenet2Take the cutting board
t3Put the cutting boardt4Take the rag
t5Put the ragt6Take the plate
t7Put the platet8Take the spoon
t9Put the spoont10Take the cup
t11Put the cupt12Leave the scene
t13Take the cutting board and the ragt14Put the cutting board and the rag
t15Take the plate and the cupt16Put the plate and the cup
t17Take the fork, knife, and spoont18Put the Fork, knife, and spoon
t19Open cupboardt20Take the fork, spoon, and knife
t21Put the fork, spoon, and knife
Table 2. Action positioning indicator (%).
Table 2. Action positioning indicator (%).
Iou 0.3 0.5 0.7
Recall35.6%37.0%24.1%118.5%17.4%10.3%18.4%9.0%10.3%1
Precision94.1%71.2%50.0%14.3%71.4%57.1%21.4%14.2%30.1%35.7%21.4%14.2%
AP 38.9% 23.9% 16.3%
Table 3. Performance evaluation table for action recognition (%).
Table 3. Performance evaluation table for action recognition (%).
NetworkAccuracyPrecisionRecallAverage
LRCN71.354.552.859.5
3D-ConvNet76.258.257.063.8
Two-StreamI3D82.079.380.480.5
Two-Stream-CBAM-I3D85.279.380.281.5
AFSD86.180.581.282.6
Ours87.083.685.485.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, S.; Zou, M.; Cao, R.; Zhao, Z.; Zeng, Q. Video Process Mining and Model Matching for Intelligent Development: Conformance Checking. Sensors 2023, 23, 3812. https://doi.org/10.3390/s23083812

AMA Style

Chen S, Zou M, Cao R, Zhao Z, Zeng Q. Video Process Mining and Model Matching for Intelligent Development: Conformance Checking. Sensors. 2023; 23(8):3812. https://doi.org/10.3390/s23083812

Chicago/Turabian Style

Chen, Shuang, Minghao Zou, Rui Cao, Ziqi Zhao, and Qingtian Zeng. 2023. "Video Process Mining and Model Matching for Intelligent Development: Conformance Checking" Sensors 23, no. 8: 3812. https://doi.org/10.3390/s23083812

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop