Next Article in Journal
Experimental and Numerical Characterization of Violent Sloshing Flows Using a Single Degree of Freedom Approach
Next Article in Special Issue
Path Tracking of a 4WIS–4WID Agricultural Machinery Based on Variable Look-Ahead Distance
Previous Article in Journal
Distributed Ledger Technologies and Their Applications: A Review
Previous Article in Special Issue
A Highly Accurate Forest Fire Prediction Model Based on an Improved Dynamic Convolutional Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Agricultural Machinery Intelligent Design System Based on Integrating Image Processing and Knowledge Reasoning

1
College of Engineering, South China Agricultural University, Guangzhou 510642, China
2
College of Urban and Rural Construction, Zhongkai University of Agriculture and Engineering, Guangzhou 510006, China
3
Foshan-Zhongke Innovation Research Institute of Intelligent Agriculture and Robotics, Foshan 528000, China
4
Department of Chemical Engineering, University of Surrey, Guildford GU2 7XH, UK
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(15), 7900; https://doi.org/10.3390/app12157900
Submission received: 14 July 2022 / Revised: 1 August 2022 / Accepted: 3 August 2022 / Published: 6 August 2022
(This article belongs to the Special Issue Advances in Robotics and Mechatronics for Agriculture)

Abstract

:
Agricultural machinery intelligence is the inevitable direction of agricultural machinery design, and the systems in these designs are important tools. In this paper, to address the problem of low processing power of traditional agricultural machinery design systems in analyzing data, such as fit, tolerance, interchangeability, and the assembly process, as well as to overcome the disadvantages of the high cost of intelligent design modules, lack of data compatibility, and inconsistency between modules, a novel agricultural machinery intelligent design system integrating image processing and knowledge reasoning is constructed. An image-processing algorithm and trigger are used to detect the feature parameters of key parts of agricultural machinery and build a virtual prototype. At the same time, a special knowledge base of agricultural machinery is constructed to analyze the test data of the virtual prototype. The results of practical application and software evaluation of third-party institutions show that the system improves the efficiency of intelligent design in key parts of agricultural machinery by approximately 20%, reduces the operation error rate of personnel by approximately 40% and the consumption of computer resources by approximately 30%, and greatly reduces the purchase cost of intelligent design systems to provide a reference for intelligent design to guide actual production.

1. Introduction

With the continuous improvement of agricultural machinery intelligent design databases, most parts can be redesigned and put into production by means of reuse design. It is the core of intelligent design and advanced manufacturing technology of key parts of agricultural machinery to build a virtual prototype with advanced technology and complete functions for production simulation and data analysis [1,2]. However, the current cost is too high for small and medium-sized agricultural machinery equipment enterprises to purchase large-scale intelligent design systems. In designing key parts of agricultural machinery that consist of numerous parts with high similarity, the calculation efficiency and the utilization rate of software function are low, and the coherence and flexibility are weak. Therefore, the construction of a virtual prototype and its unique knowledge base of key parts of agricultural machinery with strong functionality, high data abundance, low cost, and strong practicability is of great significance for the intelligent design of agricultural machinery to guide production [3,4,5].
In this work we analyze the existing intelligent design platform and research results. In terms of system functions, SolidWorks has released its commercial plug-in for collision detection. Users can select features and define the relative motion relationship between parts to obtain interference data. However, there is a lack of global data automatic detection and a real-time feedback function of multiple parts. Rameau designed the calculation method of fit clearance under various assembly conditions [6]; however, the computational efficiency of this method was low, and the algorithm process was cumbersome for the agricultural machinery assembly data with strong regularity. Schleich designed a multifunctional virtual prototype for product design data analysis [4,7,8]. Assad constructed a simulation system to estimate production data [9,10,11], but they both lack the data compatibility of simulation analysis function and multiplatform interaction. In terms of data management, Ahmed built a cross-system product data management platform [11]; Wu built an algorithm for reuse data analysis and call [12,13]; Zhang designed the reconstruction algorithm of the feature surface to calculate the part feature data [14,15]; and Bobenrieth used 2D sketches for reuse design and model retrieval [16,17]. However, there are still problems of poor real-time and simulation. In terms of interactivity, NX, Creo, and other mainstream commercial CAD software import parts and manually plan assembly through users during virtual assembly; Wang, used algorithms to construct assembly constraints to control part assembly [18,19,20,21]; and Alejandro combined virtual reality technology to develop a more interactive and practical design system [4,22,23,24].
In summary, the above research still contains the problems of low system function coherence, low simulation degree, poor interaction, high cost, low calculation efficiency, lack of dynamic update of data, and real-time analysis ability [25]. Based on the existing research, this paper further combines the features of key parts of agricultural machinery and advanced technologies, such as image processing and deep learning, to detect the feature parameters of parts and construct its virtual prototype and verifies the data accuracy of the virtual prototype by the method of the curvature radius of contour points. A special knowledge base containing enterprise production process data is constructed, a variety of design tests and data analyses are conducted using a virtual prototype, and the reliability of the knowledge base data is verified by comparing the theoretical value with the actual value. The virtual prototype and knowledge base are integrated in the virtual reality engine, and an intelligent design system for key parts of agricultural machinery is developed. Finally, a practical production case is used to indicate the practicality of the system. The main contributions and innovations of the article are as follows:
(1)
The system uses the image-processing algorithm, with higher measurement accuracy and trigger, to measure the characteristic parameters of the 3D model of parts and reconstructs the virtual prototype with the simulation appearance, collision, and interference detection function and simulation of the physical characteristics of the 3D model according to the parametric design process of mechanical parts.
(2)
The system constructs a knowledge base that can be updated dynamically, infers and updates the data of assembly sequence, assembly benchmark, fit tolerance zone and product interchangeability of the virtual prototype in real-time, and carries out production tests and analyses of the test data. At the same time, the system can freely match the knowledge reasoning program of product tests according to the design requirements, and carry out data analysis flexibly, coherently, and automatically.
(3)
According to the assembly logic of key parts of agricultural machinery, the system simplifies the assembly behaviour, uses VR equipment to complete the virtual assembly with stronger interactions and higher immersion, and uses the knowledge base to feed back the analysis results of assembly interference data and user assembly operation specifications in real-time. This helps to improve the efficiency of assembly training of small and medium-sized enterprises.
(4)
Based on the virtual reality engine, this paper constructs an intelligent design system for key parts of agricultural machinery with more powerful real-time and calculated functions. Finally, the third-party software evaluation and practical production application test results show that the system is more in line with the actual production situation of small and multi-batch, and changeable production requirements of small and medium-sized enterprises. It can effectively improve the design and production efficiency and reduce the production cost.
The structure of the article is as follows: the second chapter is the process of virtual prototype construction and accuracy verification; the third chapter is the construction of a knowledge base and the verification process of data reliability; the fourth chapter is the overall application process of the system and the analysis of a third-party software test and evaluation data; and the last chapter summarizes the advantages and disadvantages of the system.

2. Constructing Virtual Prototype

The traditional 3D model prototype obtains test data through various CAD test modules. Its essence is based on the numerical calculation and features data interaction of a single part model prototype, which lacks global real-time data feedback and an independent data calculation function. The virtual prototype constructed in this study uses the collision trigger mechanism of the mesh solid model to automatically feedback interference data and assembly data. The data storage and feedback of each part can be completed independently and synchronously through the data script, which has the advantages of real-time data calculation, independent feedback, and global synchronous analysis.
The object structure of the virtual prototype is a multilevel structure, including a mesh solid model layer for physical simulation and data acquisition, a data layer for accessing part of the CAD data, an assembly datum layer for assembly behaviour definition, and a 3D model layer for appearance rendering. The mesh solid model and data layers access and calculate data through scripts, which can be called by the main program of the system. The call logic of each part is shown in Section 3. The schematic diagram of its structure level and screenshot of the system are as shown in Figure 1.

2.1. Construction of the Mesh Solid Model Layer

To train the deep-learning network that recognizes the part category and the feature category contained in the part, we need to use the sample image data set of the model. The system needs to solve the problems of file format conversion, model size normalization, and model centroid coordinate system reconstruction of a large number of models, so that the image sample size of the model and the model attitude are both consistent. The mesh solid model layer refers to the mesh solid model composed of a collision bounding box and corresponding physical simulation components that fit the contour of the part. It has the functions of multibody continuous collision detection and triggering, simulation of physical properties, and so on. It is the basis for the system to obtain all test and analysis data. The built-in mesh bounding box attribute of the virtual reality engine only has the trigger bounding box fitting with the outer contour of the 3D model for the nonconvex 3D model but does not have the function of collision detection and the simulation of physical attributes of a rigid body. However, according to the parametric design process of mechanical parts, we can make the mesh bounding box have the attributes of a multibody continuous collision trigger and simulation physics by slightly dividing it into a set of mesh bodies regularly spliced by a limited number of convex sub-meshes [2,25].
The architecture of this section is as follows: in Section 2.1.1, the system pre-processes the 3D model of key parts of agricultural machinery and constructs the image dataset for deep-learning network training [26]; in Section 2.1.2, based on the 3D model of the part, the system uses the image-processing algorithm to detect the key feature parameters of the part from the orthogonal view image; in Section 2.1.3, the system uses the differential method to construct the general mesh solid model of parts.

2.1.1. Training of the Feature Detection Network

First, we collect approximately 10,180 CAD three-dimensional models of key parts of agricultural machinery, convert their file formats in batches using the macro command of SolidWorks and 3DMAX, and import them into the virtual reality engine. Second, combined with the structural characteristics of the rotating body of the key parts of agricultural machinery, the system constructs the cube bounding box. According to the side length relationship of the bounding box, the posture of the parts is adjusted until the rotation axis is parallel to the x-axis of the world coordinate system and scaled to the appropriate size (see Figure 2). Through experiments, the best resolution of image samples in the dataset is 960 × 960, and the best size of parts in the sampling environment is 4. According to this conclusion, the system obtains the value of the rotation axis side length (L) and radial side length (R) of the cube bounding box of the three-dimensional model of the part and scales them into R = 4, L = L × 4/R [27].
At the same time, the shaft and general parts are divided according to the proportion of radial and axial dimensions of the parts.
Then, the system samples the parts one by one in the three orthogonal view directions of the top view, front view, and side view to obtain the image features of the rotating surface of the parts in the front view and the cross-section of the parts in the side view. It should be emphasized that the centre of the sample image corresponds to the projection point of the part centroid, which is the reference point for measuring the feature parameters and is used to describe the relative position relationship of each feature on the part [28]. Finally, the system constructs the image dataset for the training of each feature detection network [27].

2.1.2. Detection of Feature Parameters

The system uses the sample image data set constructed in Section 2.1.1 and trains the deep-learning network for classifying part categories and part feature categories; at the same time, OpenCV image-processing algorithm is used to detect the feature parameters of parts, and the goal of extracting feature parameters from part sample images is completed. The essence of the virtual sample mechanism construction of parts is the data reconstruction of key features of parts. The assembly features of key parts of agricultural machinery are clear and classified. The system uses an image-processing algorithm to accurately measure the feature parameters, makes full use of the feature classification function of the deep-learning network with high precision, and avoids the accurate data measurement function of deep-learning instability. Referring to the data expression of engineering drawings, the feature data are measured from two-dimensional images, which improves the detection efficiency and accuracy.

Use of an Image Algorithm to Detect Feature Parameters

The key feature parameters of parts include the design parameters of mechanical parts such as tooth tip circle radius, number of teeth, tooth width, arc outer diameter, and arc inner diameter, for which the system sets corresponding detection means. Table 1 and Table 2 summarize the feature combination of all key parts of agricultural machinery and the key parameters of each type of feature.
The feature contour edge of the top view of the part can be summarized as a multi-segment square wave function. The edge of the contour of the part in the image sample is equivalent to the square wave function. Extracting the key feature parameters of parts can be equivalent to measuring and screening the discontinuities of function and slope. Based on the classification and positioning of the low-precision data of the feature detection network, we compare the positioning data of the image-processing algorithm with the data of the former to accurately locate and classify each feature and extract its parameters. For example, the centroid coordinate of a straight gear feature in the detection result of the feature recognition network is (500, 1440), and the system assigns its feature type result to the feature segment closest to (500, 1440) in the feature extracted by the image-processing algorithm. In addition, the unassigned feature segments are arc feature segments by default. This step avoids the positioning error of the feature recognition network (for example, all pixels of the feature are not completely selected in the right a priori box in Figure 3) and makes full use of its target classification function to judge the type of feature segment. Among them, the process of the image-processing algorithm is as follows:
It can be seen from the processing results that the feature contour edge of the top view of the part can be summarized into a multisegmented square wave function. The processing of the contour edge is actually the detection and screening of the discontinuity point of the square wave function and the band with zero or positive infinity slope.
P i = ( x i , y i ) , ( P i c a n n y , y i < H e i g t h 2 )
| y i + 1 y i | = y i
U n i t = S i z e × 1000 2880
{ y i K Y { y i + 1 = m i n { y i , y i + 1 } R j = ( H e i g h t 2 y i + 1 ) × U n i t S j = S j { P i + 1 } } y i > K Y { J = J { P i + 1 } R = R { R i + 1 } j = j + 1 } }
L j = | J j + 1 J j | × U n i t
C j = J j + 1 J j 2 × U n i t O
L i K L , { L = L L i R = R R i C = C C i }
First, the system uses the Canny algorithm to extract all contour edge points ( c a n n y ) in the upper half of the image with height ( H e i g h t ) and width of 2880 px (Formula (1)). At the same time, the conversion scale operator U n i t between pixel units and millimeters is solved by the square field size ( S i z e ) of the virtual camera (Formula (3)).
Second, based on whether the pixel coordinate height difference Δ y i of the adjacent point P i , P i + 1 exceeds the threshold K Y , the algorithm selects points P1 and P3 in Figure 4 as the starting point of the next feature contour and the end point of the current feature segment, respectively, and stores them in the set J , S j , where the number j of points set J represents the number of feature segments; Set S j stores all contour edge points of each feature segment.
At the same time, whenever the algorithm records a new segment of feature points, the minimum height of all points in Sj is used as the contour radius Rj of the current feature segment (see the white vertical line segment in Figure 4), the horizontal distance between the starting point and the last point is taken as the length Lj of the feature segment, and the midpoint of the starting point and the last point is selected to solve the relative position relationship between it and the midpoint of the image O as the centroid C j of the feature segment (see point P0 (1440, 1440) and vector V04 in Figure 4).
Finally, the algorithm filters the length Lj of each feature segment according to the threshold KL and obtains the centroid relative coordinate set C, length set L, and radius set R of all key feature segments.

Use of a Trigger to Detect Feature Parameters

The feature recognition network has poor accuracy in identifying a large number of features. The system uses the detection ball with a collision trigger function to detect the number of teeth, helix angle, rotation direction, shaft hole, internal spline, and other parameters of the parts. Among them, the radius of the unit sphere and the thickness of the detection disc are very small, and the trigger information between them and the part trigger can be returned. The schematic diagram of the test ball is as Figure 5.
R j = 0.98 × H j
θ r = 2 a r c s i n ( r R j )
P i + 1 = M x ( θ r ) P i
(1)
Process of detecting the number of teeth: the system controls the detection ball to move according to the track in Figure 6a, records the trigger times, and solves the teeth.
where Rj is the radius of the track of the detection ball. θr is the rotation angle of the detection ball per frame. Hj is the radius of the addendum circle. r is the radius of the detection ball; P i , P i + 1 is the position of the detection ball before and after rotating for one frame. Mx(θr) is the rotation matrix about the x-axis.
t a n β = R × t a n β k R k
(2)
Process of detecting the helical angle and direction of the helical gear: As shown in Figure 6b, we take the midpoint P0, P1, P2, and P3 of the adjacent trigger points in the detection ball tracks Ball1 and Ball2 on both sides and combine them into paths H1, L0, and H2. Among them, P1, P2, and P3 are obtained from the adjacent points on the left and right sides of the projection point of P0 on the ball2 track. The helix corresponding to the path without a trigger can be solved using Formula (11). β is the slope corresponding to L0, and βk is the helix angle obtained by the conversion relationship between the track radius R and the dividing circle radius Rk.
(3)
Process of detecting spiral angle and direction of spiral bevel gear: The target detection algorithm and image-processing algorithm locate and extract the maximum and minimum radius point B and point s of the conical tooth feature contour on the image and solve the radius of the tooth top circle at the large end and small end, sub cone angle, and tooth width. According to the algorithm in (2), we solve the spiral angle (see L1 and L2 in Figure 6c,d) and direction of the large end and small end of the gear.
(4)
Process of detecting the shift fork: The image-processing algorithm obtains the positions of the shift fork ring and the shift fork shaft hole from the deep-learning target detection network and obtains the fork ring attachment point P1. P2 is triggered by the movement of the detection ball along the axis of the shift fork hole, and P3 is triggered by the movement along F2 again and solves the radius and relative position of the fork ring (see Figure 7).
(5)
Process of detecting inner contour feature segments: The system controls the two detection discs, D1 and D2, to move along the rotation axis direction (the direction of the white arrow in the Figure 8) from the left end face of the part. The inner diameter of the characteristic section is detected by detecting the trigger state changes of D1 and D2. The inner contour step can be summarized into two states: state1 and state2. After D1 and D2 move to the right end face, the internal spline features are determined according to the detection process of the number of teeth (Figure 6a).

2.1.3. Mesh Solid Modelling

The key parts of agricultural machinery follow strict parametric design. This study starts with the design parameter equation of each part to fit the part contour, and the generated mesh solid structure corresponds to the real structure. For example, the involute equation is used to fit the gear. Compared with the traditional complex surface fitting algorithm, its mesh solid accuracy and calculation efficiency are higher. The mesh solid model built by the system is the core part of the virtual prototype, which solves the problem of the weak ability of the traditional virtual prototype in physical simulation and continuous collision and interference detection of multiple parts.
The arc feature can be decomposed into a mesh solid model composed of a finite number of trapezoidal elements. The gear features can be decomposed into finite tooth section elements along the axial direction to fit the gear features of helical, spiral, and conical teeth. The bright green wireframe in the following figures represents the bounding box of the part. Figure 9a is the 3D model; Figure 9b shows the built-in mesh trigger without physical simulation and collision trigger detection function; Figure 9c is the mesh solid model constructed in this paper, including the gear feature and arc feature of the inner hole.

Gear Feature Mesh Solid Modelling

R b = M × Z × c o s 20 ° / 2
φ a φ b A T × i + φ b = φ i
a r c s i n ( M × π / 4 ) z 2 + y 2 + a r c t a n y / z = δ r
t = B / A r
2 × π / Z = χ
First, the algorithm divides the rolling angle range [ φ b , φ a ] into finite values φ i according to the accuracy A T and obtains the corresponding points P i on the involute; among them, φ a is the rolling angle corresponding to the intersection P2 of the addendum circle and the involute, and φ b is the rolling angle corresponding to the intersection P1 of the base circle and the involute. Then, the involute rotates around the origin δ r radians, mirrors clone along the z-axis to obtain all differential contour points of the gear section and translates and clones the gear section points on the back to construct the mesh element (see gear section and mesh element). In Formula (12), M is the module of the spur gear and helical gear and the module of the large end of the spiral bevel gear; Z is the number of teeth of the gear; and y , z are the Y and Z coordinates of the intersection of the dividing circle and the involute. B , A r are the width of the gear and the division accuracy of the tooth line, respectively. When the gear is a cylindrical spur gear, each mesh element is arranged along the axial direction (see spur gear in Figure 10). When the gear is a cylindrical helical gear or a bevel gear, each mesh element with a thickness of t is arranged along the helix or spiral direction (see the black dotted line in helix gear and spiral bevel gear) to form the mesh solid model of a single gear tooth. Finally, the mesh solid model of a single gear tooth is cloned Z times, and each clone rotates χ radians around the rotation axis of the part to construct the mesh solid model of gear features (see helix gear and spiral bevel gear in Figure 11).

Arc Feature Mesh Solid Modelling

The algorithm divides the arc into broken segments and uses a finite number of trapezoidal mesh elements to fit the arc features. The fitting diagram and effect diagram of the mesh solid model are as Figure 12. The red trapezoid represents a mesh unit with an arc feature.
α k = k × θ A a
p i = [ x c ± T a 2 R × cos α k R × sin α k ] T
The algorithm divides the arc in which the center angle is θ into A a arc elements. p i is the vertex on the arc element, and α k is the included angle between its radius line and y-axis. The value range of k in Formula (17) is [ 0 , A a ] . When p i is on the outer contour, R is the outer diameter of the contour. In contrast, R is the inner diameter. x c , T a represents the centroid X coordinate and thickness of the arc feature.

2.1.4. System Error Analysis

Error Analysis of the Feature Parameter Detection Process

In this study, GoogLeNet is used to classify the parts corresponding to the image according to the sample image of the part; YoloV4Tiny is used to detect all part features in the sample image according to the type of features contained in the part. Their dataset is classified according to the types of parts and their characteristics summarized in Table 1. There are about 1500 samples of each type of parts. The training set, test set, and verification set are divided according to the quantity ratio of 8:1:1 to construct the data set of GoogLeNet image classification network. This data set, according to the relationship between parts and their feature types summarized in Table 1, frames the data features in various parts’ image samples, and constructs the data set of YoloV4Tiny. The schematic diagram of the data set is shown in Figure 13. Their recognition and detection accuracies are both higher than 99.6%. In this study, approximately 10,000 three-dimensional models are used to construct its training set. Some of its data and the PR curve of network training are as shown in Table 3 and Figure 13 and Figure 14.
On the other hand, the error unit of the image-processing algorithm is the unit length of pixels. We randomly selected several three-dimensional models with known part parameters, compared them with the system detection data, and solved the relative error of the data. The results are as shown in Table 4.
From the data in Table 4, it can be concluded that the relative errors of some feature parameters are below 2 × 10−5, which is better than the measurement results of the feature recognition network (see the detection results in Figure 3).

Error Analysis of Mesh Solid Modelling

The error of differential calculation will occur in the arc fitting of the mesh solid model. Increasing the values of A r , A T , A a , which are in Formulas (13), (15), and (17), can improve its contour accuracy. Using the error analysis method of curve fitting [29,30,31], we selected any vector diameter in the cylindrical coordinate system to intersect the mesh solid model and the 3D model, obtained two surface contour points ( r , θ , h ) , ( r r , θ r , h r ) , and then solved the relative error σ i of their spatial position to evaluate the fitting accuracy of the mesh solid model. The average frame rate data are used to evaluate the consumption of computing resources of mesh solid models with different division accuracies. The statistical data are shown in Table 5.
σ i = | ( r , θ , h ) ( r r , θ r , h r ) | | ( r r , θ r , h r ) |
Meshing accuracy refers to the number of differential line segments fitting the curve of unit length. When its value is larger, the fitting accuracy is higher. On the premise of ensuring that the fitting error is lower than ±0.1 mm of the machining standard, when the division accuracy is taken as 5, the fitting error is about 0.1%, equivalent to ±0.001 mm. At this time, the operating frame rate of the system is about 58 Hz; when the division accuracy is taken as 10, the reduction of fitting error is not significant, and the operating frame rate drops to 55 Hz; when the division accuracy is greater than 10, the system runs stuck.
The conclusion shows that the best division accuracy of the mesh solid model is 5. On the premise of ensuring the accuracy of the design process, the mesh solid model avoids the complex surface fitting process of traditional algorithms and constructs a virtual prototype similar to the actual structure.

2.2. Construction of the Assembly Datum Layer

The traditional virtual assembly actually uses geometric logic to describe the general assembly behaviour, which requires the operator to have high geometric language analysis and judgement ability, lacking intuition and authenticity. This paper simplifies the regular assembly behaviour of key parts of agricultural machinery. The operator uses VR for virtual assembly to improve the interactivity and authenticity. The system uses simple assembly behaviour, consistent with the actual operation, to assemble, ensuring the consistency of users’ operation and reducing the difficulty of virtual assembly.
The system uses the geometric logic of the assembly datum to define the assembly behaviour of parts. The assembly datum axis and assembly reference plane are located according to the part rotation axis and step surface of the mesh solid model, which are used for shaft hole alignment assembly and part assembly positioning, respectively, and their positioning error is the same as that of the mesh solid model. Designers screen the incorrect assembly benchmark through the visual interface to improve the fault tolerance rate of the system (see Figure 15).

2.3. Construction of the Data Layer and 3D Model Layer

The data layer is a node where each part of the data is calculated and stored independently. Multiple hash tables are set in the script of this node to store the data from the mesh solid model and knowledge base. During the assembly process, the system continuously calls the node data for knowledge reasoning and feeds back the assembly test data in real time. The sub-nodes of the 3D model layer of the virtual prototype correspond to the 3D model of the part, which improves the data processing efficiency of simultaneous feedback from multiple parts.

3. Construction and Application of Knowledge Base

The design parameters and calculation logic of the traditional intelligent design system have been encapsulated in the program, so it is impossible to change the design standards and calculation logic according to the changeable requirements of small batch production. The knowledge base built by this system is completely inputted and defined by users, and different calculation logic and design standards can be specified according to different production needs.
The system constructs the mesh solid model of a virtual prototype with various design and test functions. It can call the design process data and script logic commands in the knowledge base to control multiple parts for multiple design tests at the same time. The virtual prototype of each part can detect and feedback its own test data independently. At the same time, the knowledge base data are dynamically generated by data tables.
Compared with the traditional agricultural machinery intelligent design system, which completes various design and test functions and highly encapsulates design process data in different commercial plug-in interfaces, this system has an independent, automatic, and synchronous single part data processing function and dynamically updated user-defined design process data, which is more in line with actual production needs.
The knowledge base is divided into five parts. Among them, the knowledge base of assembly, part classification, and assembly sequence evaluation calculates the data according to the judgement logic and calculation process in the script. Tolerance fit, interference analysis, and interchangeability knowledge base need to use feature radius and accuracy grade to retrieve the table data under the root directory of the system, and then use script for operation.

3.1. Knowledge Base for Virtual Assembly

Virtual assembly controls the geometric calculation of multi-object interactions by means of dynamics or kinematics. The system carries out human–computer interactions using VR equipment, calls the mesh solid model layer for collision feedback [32,33], simplifies the assembly logic into two assembly behaviours: datum axis alignment and datum plane alignment [34], and uses the kinematic equation to describe the motion of the object. Finally, the system sets the range threshold and adjusts the part to the assembly pose [21,35].
θ = a r c t a n ( a b a p | a b | | a p | )
L p = f p f b
P = R p P + L p
In the formula, θ is the included angle of the assembly datum axis between the assembly part and assembly body; a p , a b and f p , f b are the direction vectors of the assembly datum axis and the normal vector of the assembly datum plane of the parts and assembly body, respectively; L p is the displacement vector of the part; R p is the rotation matrix rotating around a specific axis with the centroid of the part as the origin; and P , P are the position vectors before and after part assembly.

3.2. Knowledge Base for CAD

In the process of CAD, the tolerance zone, fitting form, and assembly sequence of parts are designed according to the mechanical manufacturing process standard. Then, combined with the production needs of the enterprise, the design evaluation standard evaluates the design scheme and guides the computer-aided design. Among them, the data table of the knowledge base is the .csv binary table file in the root directory of the system. The system retrieves the data of the tolerance table, dimension deviation table, sequence evaluation data table, and interchangeability data table with the parameters of feature diameter, accuracy grade, and evaluation index, inputs the corresponding script for calculation, outputs it to the system, and updates the data of the virtual prototype. It should be emphasized that every time the system starts, it traverses the .csv file and dynamically generates a hash table, and the data in the .csv file can be updated. This solves the problem that traditional intelligent design platforms face, in that it cannot change the calculation logic and production data encapsulated in the system to meet the changing design requirements of small batch production. The knowledge base structure is shown in Figure 16.

3.2.1. Knowledge Base of Part Basic Classification

This part of the knowledge base data is encapsulated in the script code. First, the system converts the 3D model into triangular patch structure according to the process of Section 2.1.1. FBX model format (see Figure 17 format conversion). Then, the system reconstructs the centroid coordinate system of the model and adjusts the zoom ratio (see Figure 17, reconstructs the centroid coordinate system) and samples according to the fixed pose of the centroid of the model to obtain the sample image (see Figure 17, sampling). Among them, the system classifies axis parts and general parts according to the length width ratio of the model bounding box after reconstructing the centroid coordinate system (see Figure 17 part classification). Finally, the system inputs the sample image of general parts to GoogLeNet to classify general parts such as gears and bearings; then inputs all classified part sample data into OpenCV image-processing algorithm to detect all feature parameters, and inputs them into YoloV4tiny to detect the feature types contained in each category of parts (see Figure 17 feature parameter extraction).

3.2.2. Knowledge Base of Assembly Sequence and Evaluation

The assembly main body of key parts of agricultural machinery is generally the shaft, and its assembly parts are stepped or spline shaft segments, corresponding to each feature on the mesh solid mode [36]. The system matches the feature with the feature parameters of the part to generate a variety of assembly sequences. This part of the knowledge framework is shown in Table 6.
Taking the assembly main body axis in the upper part in Figure 18 as an example, the algorithm takes the x-axis (red arrow) of the assembly main body as the positive direction and uses the assembly feature parameters with frames A and B to match the parts. For example, the assembly serial numbers of Parts C1, D1, D2, E1, and E2 are (+2.1), (+5.1), (+5.2), (+7.1), and (+7.2), respectively, where D2 (+5.2) represents the part of the second layer of the fifth part combination of forward assembly (Box D represents a part combination).
We count the actual production data of the workshop, such as the maximum and minimum values of the assembly time, the moving distance of the assembly personnel in the workshop, and the replacement frequency of assembly tools and other indicators. Then, we divide the score range of 100 with a step of 10 points and store it in the knowledge base. The system calls the assembly index values of various parts within this framework to calculate the evaluation value of each assembly sequence [23]. Part of the data of the knowledge base is shown in Table 7.
i × 100 k j = α i j
T k = α j
Formula (23) solves the assembly time and assembly difficulty evaluation value α i j of the ith level of the jth evaluation index of the assembly sequence. k j is the number of index division levels. Formula (24) is the total evaluation value of the kth assembly sequence.

3.2.3. Knowledge Base of the Tolerance Fit and Interference Analysis

The interference amount to be tested in the tolerance fit design is mainly at the shaft hole fit, including the interference fit amount and interference amount caused by improper design. The user sets the tolerance grade of the part and retrieves the tolerance zone data according to the feature diameter. Then, the system solves the size deviation and limit size of the matched part features according to the set matching benchmark system. Next, the system updates the parameters of the solid model layer of the part mesh to result in dimensional deviation. Finally, when the user carries out virtual assembly, the system retrieves the data within the knowledge framework according to the diameter of the feature and the set accuracy level to analyze whether the interference of the part is reasonable in real time. Part of the knowledge framework is shown in Table 8.
The system transforms the color of the mesh cell body in the part mesh solid model, marks the interference position, extracts its centroid, and produces the scatter diagram of interference distribution, which intuitively reflects the distribution of interference quantity, as shown in Figure 19.
η i = m i n { R r ( y i j y i c ) 2 + ( z i j z i c ) 2 }
η y z = m a x { | η i | }
η x = | | x c 1 x c 2 | ( T 1 + T 2 2 ) |
( x i j , y i j , z i j ) is the centroid coordinates of the jth mesh cell on the assembly part feature that interferes with the ith mesh cell on the feature of the assembly body. ( x i c , y i c , z i c ) is the centroid of the ith mesh cell body in which interference occurs on the feature of the assembly body. η i is the interference value of the ith mesh cell on the feature of the assembly body; when it is negative, the smaller the value is, the greater the interference; when this value is positive, there is no interference. Take the maximum value of all mesh elements that interfere with the assembly body as the radial interference η y z . The axial interference η x is calculated by the distance between the centroid X coordinate x c 1 , x c 2 of the interference feature and the feature thickness T 1 , T 2 .

3.2.4. Knowledge Base of the Parts Interchangeability

The production process statistics of multiple workshops within this knowledge framework is used to calculate the matching interference law of different batches of parts and generate the interchangeability matching scheme of products in each workshop [7,15], which is used to estimate the production cost and modify the tolerance matching scheme. For application examples, see Section 4.2.

3.2.5. Fault Tolerant Processing of the System

In the process of knowledge reasoning and building of a solid mesh model, the following errors will occur with low probability: (1) feature recognition error; (2) assembly sequence generation error. The system sets the mesh solid model visualization interface and assembly sequence visualization interface, on which the setter can view the speculation results and manually change the data. Users can reconstruct the mesh solid model by modifying the script parameters mounted on the mesh solid model layer, as well as modify the assembly sequence by modifying the assembly sequence number in the data layer of each part.

4. System Application Example

Using the bearing selection scheme data of transmission components of an agricultural machinery gearbox, designed and produced by an enterprise, the recommended bearing model is generated after automatically completing the construction of the virtual prototype, and the accuracy of the system is verified by comparison with the actual process data. Finally, the system performance is evaluated by the software evaluation of the third-party organization. The field data measurement is shown in Figure 20.

4.1. Construction of Virtual Prototype and Generation of Assembly Sequence

The assembly sequence visualization interface and the assembly sequence number of each part are shown in Figure 21, as well as the assembly sequence evaluation data in Table 9.
The reason why Plan B is chosen is that the last part of the right half and the first part of the left half needs to be installed with a tail hook wrench. If the right half is installed first and then the left half is installed, the time of repeated tool access will be saved.

4.2. Generation of Assembly Interference and Matching Schemes and Incomplete Interchange Schemes

We investigated and obtained the production process statistics of workshops I and II (Table 10), updated the mesh solid model layer of the virtual prototype for the updated data statistics (Table 11), and counted the interference data and interchangeability matching data after virtual assembly (Table 12).
According to the data in Table 12, the product size of workshop 1 is small and that of workshop 2 is large. The system suggests that the bearing with a larger mean limit size should be matched with the product of workshop 1, while the bearing with a smaller mean value should be matched with the product of workshop 2. Then, according to the given data, we select three types of bearings in the table for virtual assembly. Finally, the interference data of the system statistics show that it meets the proposed fit clearance requirements.

4.3. System Performance Evaluation

The cooperative enterprise of the project funded by this paper employs a third-party evaluation company to conduct software evaluation. Using a computer configured with 32 GB memory, a CPU of interi59400f, and graphics card of NVIDIA 2070 Super as the test platform, the time and assembly operation error rate of four professional evaluation personnel (No. 1–4) completing the intelligent design task of this chapter, using the intelligent design system of SolidWorks and this system are counted as the performance indicators of this system. Each professional completes the design step and assembly step five times (see Figure 22). We count the average operation time, assembly error rate, and computer resource consumption rate of each person; the results are shown in Figure 23.
From the evaluation data, the design efficiency of this system is improved by approximately 20% compared with that on the SolidWorks platform, the operation error rate is reduced by approximately 40%, and the resource consumption of CPU and ram is reduced by approximately 30%. Moreover, the cost of using the commercial version of SolidWorks and its plug-in function package is approximately 200 thousand yuan, which is much higher than the selling price of the commercial development platform Unity3DPro used in this system.

4.4. Conclusions

The system can generate the analysis data required for production design and can complete the data analysis of the complete assembly scheme and interchangeability scheme, automatically and consistently. On the premise of ensuring that the design data meet the design accuracy, the system generates a preliminary design scheme reference meeting the production process requirements. In the early stages of part design, the system can use limited data to generate a more comprehensive design scheme, improve the production design efficiency of small batches and multiple changes, reduce the error rate of manual operation, and reduce the cost of purchasing intelligent design software and computer hardware configuration.

5. Summary

This research combines deep learning and image-processing technology to build a virtual prototype with complete functions and a novel structure. Combined with knowledge reasoning and data-driven technology, a special knowledge base is constructed, which is integrated into a flexible, fully functional, and practical agricultural machinery intelligent design system in a virtual reality engine.
(1)
Based on the existing 3D model, this study automatically batch constructs a general virtual prototype of key parts of agricultural machinery, integrating the functions of interference detection, cooperation analysis, and physical simulation. In the early stages of design for small and medium-sized enterprises, the limited existing design data can be reused, and new parameter schemes can be obtained according to the new production process, which can significantly reduce the production cost during the early stages of reuse design. The data volume of the lifting system can also be extended to the fields of automobiles, aircraft parts, and so on.
(2)
The system knowledge base takes the dynamically updated tabular data as the knowledge framework for knowledge reasoning; additionally, its flexibility is much higher than the product testing function encapsulated in traditional CAD software. Through the combination of various functions of knowledge reasoning, the intelligent design process of various parts is completed. Compared with the traditional CAD software, it obtains various analysis data by calling various plug-in interfaces and then integrates them into a new design scheme, which has higher efficiency and a more coherent design process. Each plug-in of traditional CAD software needs to be purchased separately, while this system is based on a virtual reality engine to freely integrate a variety of functions with lower cost.
(3)
With the help of a data visualization interface, the system provides an operation interface for designers to modify the wrong mesh solid model feature parameters and screen knowledge reasoning data to improve the fault tolerance of the system. At the same time, the feature recognition network can be used to extract more useful data from CAD drawings of reuse design.
(4)
The system is oriented to the production application of small batch and changeable processes in small and medium-sized agricultural machinery equipment enterprises and overcomes the disadvantages of low development freedom, high cost of module function integration, and low module compatibility of large-scale intelligent design CAD systems.
There are still deficiencies in the accuracy of dynamic simulation data and finite element analysis function of this system; the finite element function and precision matrix operation program library will be further integrated in the future. Additionally, we will learn from the functions of traditional CAD intelligent design systems to further improve the practicality of the system.

Author Contributions

Conceptualization and methodology, X.Z. and C.L.; validation, C.L., P.Z. and J.L.; data curation, C.L.; writing—original draft preparation, C.L., Y.T., P.Z., G.L. and J.L.; writing—review and editing, X.Z., G.L. and Y.P.; visualization, C.L. and G.L.; supervision, X.Z.; funding acquisition, Y.P., G.L. and X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by The National Key Research and Development Program of China (2017YFD0700100) and Provincial Science and Technology Department Funding Project of Jiangxi, China (Grant No.20212ABC03A27).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We sincerely acknowledge the project funding support. We are also grateful for the efforts of all our colleagues.

Conflicts of Interest

The authors declare that they have no conflict of interest.

References

  1. Yepez, P.; Alsayyed, B.; Ahmad, R. Automated maintenance plan generation based on CAD model feature recognition. Procedia CIRP 2018, 70, 35–40. [Google Scholar] [CrossRef]
  2. Zhang, Z.; Jaiswal, P.; Rai, R. FeatureNet: Machining feature recognition based on 3D convolution neural network. Comput. Aided Des. 2018, 101, 12–22. [Google Scholar] [CrossRef]
  3. Fletcher, C.; Ritchie, J.; Lim, T.; Sung, R. The development of an integrated haptic VR machining environment for the automatic generation of process plans. Comput. Ind. 2013, 64, 1045–1060. [Google Scholar] [CrossRef]
  4. Hu, L.; Liu, Z.; Tan, J. A VR simulation framework integrated with multisource CAE analysis data for mechanical equipment working process. Comput. Ind. 2018, 97, 85–96. [Google Scholar] [CrossRef]
  5. Laseinde, O.T.; Adejuyigbe, S.B. Design of plastic spur gears using virtual reality. Int. J. Comput. Aided Eng. Technol. 2014, 6, 48–61. [Google Scholar] [CrossRef]
  6. Rameau, J.F.; Serré, P.; Moinet, M. Clearance vs. tolerance for rigid overconstrained assemblies. Comput. Aided Des. 2018, 97, 27–40. [Google Scholar] [CrossRef] [Green Version]
  7. Schleich, B.; Wartzack, S. Novel approaches for the assembly simulation of rigid skin model shapes in tolerance analysis. Comput. Aided Des. 2018, 101, 1–11. [Google Scholar] [CrossRef]
  8. Rupp, S.; Müller, R. Worker assistance systems in the automotive prototype assembly–A case study. Procedia Manuf. 2021, 55, 350–357. [Google Scholar] [CrossRef]
  9. Assad, F.; Konstantinov, S.; Rushforth, E.J.; Vera, D.A.; Harrison, R. Virtual engineering in the support of sustainable assembly systems. Procedia CIRP 2021, 97, 367–372. [Google Scholar] [CrossRef]
  10. Abidi, M.H.; Al-Ahmari, A.; Ahmad, A.; Ameen, W.; Alkhalefah, H. Assessment of virtual reality-based manufacturing assembly training system. Int. J. Adv. Manuf. Technol. 2019, 105, 3743–3759. [Google Scholar] [CrossRef]
  11. Ahmed, F.; Han, S. Interoperability of product and manufacturing information using ontology. Concurr. Eng. 2015, 23, 265–278. [Google Scholar] [CrossRef]
  12. Wu, Z.; Liu, H.; Goh, M. Knowledge recommender system for complex product development using ontology and vector space model. Concurr. Eng. 2019, 27, 347–360. [Google Scholar] [CrossRef]
  13. Jaiswal, P.; Huang, J.; Rai, R. Assembly-based conceptual 3D modeling with unlabeled components using probabilistic factor graph. Comput. Aided Des. 2016, 74, 45–54. [Google Scholar] [CrossRef]
  14. Zhang, D.; Wang, X.; Hu, J.; Qin, H. Interactive modeling of complex geometric details based on empirical mode decomposition for multi-scale 3D shapes. Comput. Aided Des. 2017, 87, 1–10. [Google Scholar] [CrossRef] [Green Version]
  15. Sun, Q.; Zhao, B.; Liu, X.; Mu, X.; Zhang, Y. Assembling deviation estimation based on the real mating status of assembly. Comput. Aided Des. 2019, 115, 244–255. [Google Scholar] [CrossRef]
  16. Bobenrieth, C.; Cordier, F.; Habibi, A.; Seo, H. Descriptive: Interactive 3D shape modeling from a single descriptive sketch. Comput. Aided Des. 2020, 128, 102904. [Google Scholar] [CrossRef]
  17. Ban, S.; Hyun, K.H. 3D computational sketch synthesis framework: Assisting design exploration through generating variations of user input sketch and interactive 3D model reconstruction. Comput. Aided Des. 2020, 120, 102789. [Google Scholar] [CrossRef]
  18. Wang, Y.; Liu, J.H.; Li, L.S. Assembly sequences merging based on assembly unit partitioning. Int. J. Adv. Manuf. Technol. 2009, 45, 808–820. [Google Scholar] [CrossRef]
  19. Masehian, E.; Ghandi, S. ASPPR: A new assembly sequence and path planner/replanner for monotone and nonmonotone assembly planning. Comput. Aided Des. 2020, 123, 102828. [Google Scholar] [CrossRef]
  20. Kim, Y.; Lee, H.; Safdar, M.; Jauhar, T.A.; Han, S. Exchange of parametric assembly models based on neutral assembly constraints. Concurr. Eng. 2019, 27, 285–294. [Google Scholar] [CrossRef]
  21. Li, Z.; Wang, J.; Anwar, M.S.; Zheng, Z. An efficient method for generating assembly precedence constraints on 3D models based on a block sequence structure. Comput. Aided Des. 2020, 118, 102773. [Google Scholar] [CrossRef]
  22. Alejandro Huerta-Torruco, V.; Hernández-Uribe, Ó.; Adriana Cárdenas-Robledo, L.; Amir Rodríguez-Olivares, N. Effectiveness of virtual reality in discrete event simulation models for manufacturing systems. Comput. Ind. Eng. 2022, 168, 108079. [Google Scholar] [CrossRef]
  23. Freeman, I.; Salmon, J.; Coburn, J. A bi-directional interface for improved interaction with engineering models in virtual reality design reviews. Int. J. Interact. Des. Manuf. (IJIDeM) 2018, 12, 549–560. [Google Scholar] [CrossRef]
  24. Liu, F.; Lu, Y.; Zhang, S. A pragmatic system to support virtual assembly for military armored vehicle integrated transmission system in the virtual environment. Wirel. Pers. Commun. 2018, 102, 1337–1354. [Google Scholar] [CrossRef]
  25. Zhang, Z.; Jimack, P.K.; Wang, H. MeshingNet3D: Efficient generation of adapted tetrahedral meshes for computational mechanics. Adv. Eng. Softw. 2021, 157–158, 103021. [Google Scholar] [CrossRef]
  26. Tian, Y.; Wang, X.; Shen, Y.; Guo, Z.; Wang, Z.; Wang, F.Y. Parallel point clouds: Hybrid point cloud generation and 3D model enhancement via virtual–real integration. Remote Sens. 2021, 13, 2868. [Google Scholar] [CrossRef]
  27. Li, C.; Zou, X.; Lian, G.; Lin, J.; Zhang, P.; Chen, M.; Ye, L. Construction method of classification and detection network data set of agricultural machinery virtual assembly. J. South China Agric. Univ. 2021, 42, 117–125. [Google Scholar] [CrossRef]
  28. Liu, L.; Shi, Z.; Pan, B.; Zhang, N.; Luo, H.; Lan, X. Multiscale deep spatial feature extraction using virtual RGB image for hyperspectral imagery classification. Remote Sens. 2020, 12, 280. [Google Scholar] [CrossRef] [Green Version]
  29. Feng, C.; Taguchi, Y. FasTFit: A fast T-spline fitting algorithm. Comput. Aided Des. 2017, 92, 11–21. [Google Scholar] [CrossRef]
  30. Li, C.J.; Xie, L.L.; Du, W.B. Curve and surface fitting based on the nonhomogeneous linear differential system. Graph. Models 2019, 103, 101026. [Google Scholar] [CrossRef]
  31. Li, C.; Zou, X.; Zeng, Z.; He, J.; Li, H.; Huang, Z. Mesh entity construction algorithm of spiral bevel gear based on virtual collision body. J. Syst. Simul. 2021, 33, 837–844. [Google Scholar] [CrossRef]
  32. Kloiber, S.; Settgast, V.; Schinko, C.; Weinzerl, M.; Fritz, J.; Schreck, T.; Preiner, R. Immersive analysis of user motion in VR applications. Vis. Comput. 2020, 36, 1937–1949. [Google Scholar] [CrossRef]
  33. Brookes, J.; Warburton, M.; Alghadier, M.; Mon-Williams, M.; Mushtaq, F. Studying human behavior with virtual reality: The unity experiment framework. Behav. Res. Methods 2020, 52, 455–463. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Jayasekera, R.D.M.D.; Xu, X. Assembly validation in virtual reality—A demonstrative case. Int. J. Adv. Manuf. Technol. 2019, 105, 3579–3592. [Google Scholar] [CrossRef]
  35. Lisowski, J. Game control methods comparison when avoiding collisions with multiple objects using radar remote sensing. Remote Sens. 2020, 12, 1573. [Google Scholar] [CrossRef]
  36. Zhang, N.; Qi, T.; Zhao, Y. Real-time learning and recognition of assembly activities based on virtual reality demonstration. Sensors 2021, 21, 6201. [Google Scholar] [CrossRef]
Figure 1. Virtual prototype.
Figure 1. Virtual prototype.
Applsci 12 07900 g001
Figure 2. Cube bounding box and its datum axis.
Figure 2. Cube bounding box and its datum axis.
Applsci 12 07900 g002
Figure 3. Feature recognition network classification feature.
Figure 3. Feature recognition network classification feature.
Applsci 12 07900 g003
Figure 4. Accurate measurement of characteristic parameters by the image-processing algorithm.
Figure 4. Accurate measurement of characteristic parameters by the image-processing algorithm.
Applsci 12 07900 g004
Figure 5. Detection ball and detection discs.
Figure 5. Detection ball and detection discs.
Applsci 12 07900 g005
Figure 6. Gear feature parameter detection.
Figure 6. Gear feature parameter detection.
Applsci 12 07900 g006
Figure 7. The mesh solid model and the detection sequence of the shifting fork.
Figure 7. The mesh solid model and the detection sequence of the shifting fork.
Applsci 12 07900 g007
Figure 8. Process and mesh solid model for detecting the inner contour.
Figure 8. Process and mesh solid model for detecting the inner contour.
Applsci 12 07900 g008
Figure 9. Relationship diagram between the 3D model and the mesh solid model.
Figure 9. Relationship diagram between the 3D model and the mesh solid model.
Applsci 12 07900 g009
Figure 10. Process of mesh solid modelling of gear.
Figure 10. Process of mesh solid modelling of gear.
Applsci 12 07900 g010
Figure 11. Mesh element and mesh solid model.
Figure 11. Mesh element and mesh solid model.
Applsci 12 07900 g011
Figure 12. Arc feature mesh solid model.
Figure 12. Arc feature mesh solid model.
Applsci 12 07900 g012
Figure 13. The training dataset of the deep-learning network.
Figure 13. The training dataset of the deep-learning network.
Applsci 12 07900 g013
Figure 14. The loss and accuracy curve of the deep-learning network.
Figure 14. The loss and accuracy curve of the deep-learning network.
Applsci 12 07900 g014
Figure 15. Assembly datum axis and face.
Figure 15. Assembly datum axis and face.
Applsci 12 07900 g015
Figure 16. Data processing of the knowledge base.
Figure 16. Data processing of the knowledge base.
Applsci 12 07900 g016
Figure 17. Technology roadmap.
Figure 17. Technology roadmap.
Applsci 12 07900 g017
Figure 18. Assembly sequence diagram.
Figure 18. Assembly sequence diagram.
Applsci 12 07900 g018
Figure 19. Interference data. (a) Schematic diagram of grid interference. (b) Shaft hole interference. (c) Surface interference. (d) Interference point cloud data.
Figure 19. Interference data. (a) Schematic diagram of grid interference. (b) Shaft hole interference. (c) Surface interference. (d) Interference point cloud data.
Applsci 12 07900 g019
Figure 20. Measurement process of some parts.
Figure 20. Measurement process of some parts.
Applsci 12 07900 g020
Figure 21. Solid model of the sub−assembly part mesh.
Figure 21. Solid model of the sub−assembly part mesh.
Applsci 12 07900 g021
Figure 22. Screenshot of the operation process.
Figure 22. Screenshot of the operation process.
Applsci 12 07900 g022
Figure 23. System performance comparison data.
Figure 23. System performance comparison data.
Applsci 12 07900 g023
Table 1. Combination relationship between parts and various features.
Table 1. Combination relationship between parts and various features.
Part TypeFeature Combination Relationship
Cylindrical spur gearStraight gear feature, Ring feature of inner hole, Internal spline feature
Cylindrical helical gearHelical gear feature, Ring feature of inner hole, Internal spline feature
Spiral bevel gearSpiral bevel gear feature, Ring feature of inner hole, Internal spline feature
Shift forkRing of fork ring, Ring feature of shaft hole
shaftRing feature, Straight gear feature, Spline feature, Helical gear feature
Meshing sleeveRing feature, Internal spline feature
axle sleeveRing feature, Internal spline feature
BearingRing feature
Table 2. The combination relationship between part features and feature parameters.
Table 2. The combination relationship between part features and feature parameters.
Feature TypeFeature Parameter
Straight gear featureRadius of addendum circle, Number of teeth, Tooth width
Helical gear featureRadius of addendum circle, Number of teeth, Tooth width, Rotation direction, Helix angle
Spiral bevel gear featureRadius of addendum circle, Number of teeth, Tooth width, Rotation direction, Helix angle, Large end modulus, Small end modulus, Cone angle
Ring of forkSector angle, Inner diameter, Outer diameter, Thickness
Spline featureRadius of addendum circle, Number of teeth, Tooth width
Ring featureInner diameter, Outer diameter, Thickness
Table 3. Deep learning network training data table.
Table 3. Deep learning network training data table.
Deep Learning NetPreciousRecall
GoogLeNet99.61%99.34%
YoloV4Tiny99.23%99.87%
Table 4. Statistical table of error verification data.
Table 4. Statistical table of error verification data.
Part NamePart Feature ParametersTheoretical ValueDetection ValueRelative Error
Second driven gearRadius of addendum circle4948.99976.12 × 10−6
Tooth width2626.00031.15 × 10−5
Centroid of feature segment(−68.6, 0)(−68.5998, 0)2.92 × 10−6
Inner arc radius2929.00082.76 × 10−5
Inner arc thickness3938.99985.13 × 10−6
Centroid of inner contour feature segment(−68.6, 0)(−68.5999, 0)1.46 × 10−6
Inner bore bearing of fourth driven gearExternal diameter3737.00041.08 × 10−5
Internal diameter2524.99994.00 × 10−6
Width3030.00051.67 × 10−5
Centroid of feature segment(6,0)(6.0001,0)1.67 × 10−5
Table 5. Fitting error data of the part mesh solid model.
Table 5. Fitting error data of the part mesh solid model.
Assembly Part NameAssembly TypeMesh Generation Accuracy of Assembly PartsAverage Frame Rate
(Hz)
Mean Value of Relative Error of Mesh Solid Modelling
Third gear driving gear and input shaftFit of internal spline558.430.1368%
1055.380.1296%
Second driven gear and bearingShaft-bore fit557.960.0821%
1055.100.0857%
2nd driven gear bearing and output shaftShaft-bore fit1551.910.0596%
2048.720.0631%
Table 6. Part feature matching knowledge framework.
Table 6. Part feature matching knowledge framework.
Assembly Body FeatureCombination of Multilevel Assembly PartsMatching Basis
Spline segmentInternal spline gear-Meshing sleeve
Internal spline gear
Snap ring
Modulus
Arc segmentBearing/Shaft sleeve
Bearing/Shaft sleeve-Gear
Snap ring
Diameter
Fork slotMesh sleeve-Shifting forkDiameter
Table 7. Assembly sequence evaluation knowledge framework.
Table 7. Assembly sequence evaluation knowledge framework.
Part TypeInstallation ToolsAssembly TimeAssembly Difficulty
Driven geargreased3e3
Driving geargreased3e3
Meshing sleevegrease, screwdriverd4e4
Shaft assembly subassemblylifting armd5e5
Table 8. Tolerance zone.
Table 8. Tolerance zone.
ClassIT4
(μm)
IT5
(μm)
IT6
(μm)
Diameter (mm)
φ18–φ306913
φ30–φ5071116
φ50–φ8081319
Table 9. System generated assembly sequence scheme.
Table 9. System generated assembly sequence scheme.
Assembly Sequence PlanScoreOptimal Scheme
Plan A: +1.1, +2.1, −1.1, −2.1, −3.1, −3.2, −4.1, −4.252
Plan B: −1.1, −2.1, −3.1, −3.2, −4.1, −4.2, +1.1, +2.146
Table 10. Production process statistics table in the knowledge base.
Table 10. Production process statistics table in the knowledge base.
Workshop IWorkshop II
Accuracy ClassIT6IT7IT6IT7
Dimensional Tolerance (mm)
φ30–φ5014231625
φ50–φ8017281930
φ80–φ12020332235
Maximum allowable fit clearance10
Maximum allowable fit interference10
Table 11. Fit tolerance data of key parts.
Table 11. Fit tolerance data of key parts.
Part Features Name Tolerance Zone Basis Fit Systems Limit Size (mm)
φ110 Case input shaft locating bearing holeJ7shaft-basis system of fitsD min = 109.9825
D max = 110.0175
φ50 Input axle box body positioning bearing journalh6hole-basis system of fitsd min = 49.984
d max = 50.000
φ110 Locating bearing hole of secondary transmission axle boxJ7shaft-basis system of fitsD min = 109.9825
D max = 110.0175
φ70 Positioning bearing journal of secondary transmission axle boxh6hole-basis system of fitsd min = 69.981
d max = 70.000
Table 12. Product scheme data.
Table 12. Product scheme data.
Part NumberAverage Size of the Workshop 1Workshop 1 Average
Interference/Clearance (mm)
Average Size of the Workshop 2Workshop 2 Average
Interference/Clearance (mm)
Recommended Bearing Parameters
φ110H1φ109.9877−0.0055φ109.9867−0.008731310P0
φ50S1φ49.99350.0021φ49.99360.0025
φ110H2φ109.9970−0.0053φ109.9864−0.009032014P0
φ70S1φ69.98810.0024φ69.99240.0031
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, C.; Tang, Y.; Zou, X.; Zhang, P.; Lin, J.; Lian, G.; Pan, Y. A Novel Agricultural Machinery Intelligent Design System Based on Integrating Image Processing and Knowledge Reasoning. Appl. Sci. 2022, 12, 7900. https://doi.org/10.3390/app12157900

AMA Style

Li C, Tang Y, Zou X, Zhang P, Lin J, Lian G, Pan Y. A Novel Agricultural Machinery Intelligent Design System Based on Integrating Image Processing and Knowledge Reasoning. Applied Sciences. 2022; 12(15):7900. https://doi.org/10.3390/app12157900

Chicago/Turabian Style

Li, Cheng’en, Yunchao Tang, Xiangjun Zou, Po Zhang, Junqiang Lin, Guoping Lian, and Yaoqiang Pan. 2022. "A Novel Agricultural Machinery Intelligent Design System Based on Integrating Image Processing and Knowledge Reasoning" Applied Sciences 12, no. 15: 7900. https://doi.org/10.3390/app12157900

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop