Next Article in Journal
Hardware-In-the-Loop Simulation of Time-Delayed Anti-Swing Controller for Quadrotor with Suspended Load
Next Article in Special Issue
Studies to Overcome Brain–Computer Interface Challenges
Previous Article in Journal
Optimal Design of an Inverter-Fed PMSM for a Brake System Considering Variation in Motor Control Parameters and Input Voltage
Previous Article in Special Issue
Patent Analysis Using Bayesian Data Analysis and Network Modeling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Growing Neural Gas with Different Topologies for 3D Space Perception

Graduate School of Natural Science and Technology, Okayama University, 3-1-1 Tsushima-Naka, Kita-Ku, Okayama 700-8530, Japan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(3), 1705; https://doi.org/10.3390/app12031705
Submission received: 4 January 2022 / Revised: 27 January 2022 / Accepted: 3 February 2022 / Published: 7 February 2022
(This article belongs to the Special Issue Advances in Intelligent Systems)

Abstract

:
Three-dimensional space perception is one of the most important capabilities for an autonomous mobile robot in order to operate a task in an unknown environment adaptively since the autonomous robot needs to detect the target object and estimate the 3D pose of the target object for performing given tasks efficiently. After the 3D point cloud is measured by an RGB-D camera, the autonomous robot needs to reconstruct a structure from the 3D point cloud with color information according to the given tasks since the point cloud is unstructured data. For reconstructing the unstructured point cloud, growing neural gas (GNG) based methods have been utilized in many research studies since GNG can learn the data distribution of the point cloud appropriately. However, the conventional GNG based methods have unsolved problems about the scalability and multi-viewpoint clustering. In this paper, therefore, we propose growing neural gas with different topologies (GNG-DT) as a new topological structure learning method for solving the problems. GNG-DT has multiple topologies of each property, while the conventional GNG method has a single topology of the input vector. In addition, the distance measurement in the winner node selection uses only the position information for preserving the environmental space of the point cloud. Next, we show several experimental results of the proposed method using simulation and RGB-D datasets measured by Kinect. In these experiments, we verified that our proposed method almost outperforms the other methods from the viewpoint of the quantization and clustering errors. Finally, we summarize our proposed method and discuss the future direction on this research.

1. Introduction

Three-dimensional space perception is one of the most important capabilities for an autonomous mobile robot in order to operate a task in an unknown environment adaptively since the autonomous robot needs to detect the target object and estimate the 3D pose of the target object for performing given tasks efficiently [1,2,3,4,5,6,7]. In the research field of 3D space perception, many research studies use an RGB-D camera, such as Microsoft Kinect [8] and Intel Realsense [9], that can measure the color and depth of an image simultaneously in real time. After the 3D point cloud is measured, the autonomous robot needs to reconstruct a structure from the 3D point cloud with color information according to the given tasks since the point cloud is unstructured data.
For realizing the 3D space perception, many kinds of 3D image processing based methods have been proposed. These methods were expanded to the 3D point cloud by utilizing camera image processing technologies, such as filtering methods [10], feature extractions [11,12], and sample consensus methods [13]. However, these methods depend on feature or model descriptors, which have a problem with adaptability to environmental changes, such as geometry, texture and light conditions. To improve the adaptability, recently, the research studies of the information extraction from the point cloud were developed by a deep neural network (DNN) [14]. In particular, many research works in autonomous robot and automatic driving have proposed a DNN-based semantic segmentation method (e.g., PointNet [15], PointCNN [16] and VoxNet [17]) that can label and give the point cloud meaning. The DNN-based methods give highly accurate segmentation results by using big data with the teaching signal. However, one of the problems of the supervised learning method is application to unknown data. If the learned network is applied to the unknown environment or label not including the output layer, the network fails to segment the point cloud. In the environment where the autonomous robots are expected to perform the task, the autonomous robot needs to recognize the unknown target object and environment. Therefore, the point cloud processing method without prior knowledge is required for realizing the robot that can perform the given tasks in the unknown environment.
Self-organization map (SOM) [18] based methods are one of the main streams based on unsupervised learning for the point cloud processing [19]. The other methods that are well known in this field are neural gas (NG) [20], growing cell structure (GCS) [21], and growing neural gas [22]. Basically, these methods are called unsupervised learning when applied to an unknown data distribution without teaching signals and using the competitive learning method based on the winner-take-all approach. The SOM can project the input vectors in a high dimensional space to a topological structure in a low dimensional space, according to the data distribution. The SOM is easy to apply to many problems by determining the number of nodes and topologies, while it is difficult to design these things to the unknown data distribution. NG does not need to determine the number of nodes since the NG has node addition and deletion in the learning algorithm. However, the topological structure is updated according to the order of the data input. On the other hand, GCS and GNG can dynamically change the topological structure based on the adjacency relation held in the referred node. GCS cannot delete the nodes and edges, while GNG can dynamically delete the nodes and edges by using the concept of the edge’s age. Figure 1 shows an example result of each method applied to a 2D simulation point cloud. There are redundant nodes and edges in SOM and GCS, while GNG can preserve the 2D point cloud space appropriately. In addition, GNG can cluster the unknown ring data (Rings A, B and C) by utilizing the topological structure since there is no adjacent relations between each ring (Figure 1e). Moreover, GNG can perform noise reduction [23,24], 3D reconstruction [25,26,27,28] and feature extraction using the topological structure [29,30]. From these reasons, GNG is expected to utilize the unified perceptual system for the point cloud processing [31,32,33,34,35,36,37,38,39]. D. Viejo et al. applied GNG-based 3D feature extraction and the matching method to 3D object recognition [40]. This method just utilizes the node set for extracting features such as SHOT and spin image, which do not utilize the topological structure generated by GNG. Refs. [41,42,43,44] reduced the calculation cost and improved the adaptability of the time-series data for realizing gesture recognition and 3D object-tracking methods in real time. Ref. [41] applied a uniform grid structure to GNG for reducing the calculation cost of the winner node selection and realized the real-time 3D object-tracking method. Refs. [43,44] proposed the criteria for node addition and deletion. These methods are based on the probability density distribution of the data and nodes and verified the real-time adaptability for the point cloud data.
In this way, many kinds of GNG-based methods have been proposed from the viewpoint of various research directions. However, the GNG-based perceptual system has unsolved problems. One of the problems is that GNG cannot preserve the 3D position space if the input vector is composed of not only the 3D point cloud (position information), but also additional information, such as RGB-D data (3D position with color information). Figure 2 represents a learning example using GNG from a 3D point cloud (x, y, z) with color information (R, G, B). The generated topological structure shown in Figure 2b cannot preserve the 3D position space of the point cloud since the topological structure is generated from both the position and color information, and the scale of the color information is a dominant scale compared with the scale of the position information in this dataset. In this way, GNG-based point cloud processing methods have the scalability problem between each feature vector if the input vector is composed of multiple properties.
For solving this problem, there exist two main approaches. One is to utilize a background subtraction algorithm as preprocessing for extracting the target data and generating the topological structure of only the target object [45,46]. The other is to apply a weighted distance measurement in the winner node selection according to the importance of each property [47,48]. In the former approach, Angelopoulou, et al. proposed a background detection method based on mixture Gaussian distribution and CIE lab color space for generating topological structure of the human face and arm [45]. However, this kind of approach cannot preserve the appropriate position space if the background detection fails because of the sensor noises. In addition, the background detection needs detailed prior knowledge about the target object. Therefore, it is difficult to apply the background detection method to an unknown dataset. In the latter approach, Tunnermann, et al. proposed the GNG attention method (GNGA) for extracting saliency from a 2D camera image [47]. GNGA uses the weighted distance measurement used in the selection of the 1st and 2nd winner nodes for generating the topological structure of the position space with color information. In addition, the authors also proposed the weight-vector-based GNG method, called modified GNG with utility (GNG-U II) [48]. These methods can preserve the 2D/3D position spaces with color information by designing the appropriate weight.
Figure 2c shows an example result of GNG with the weighted distance measurement. By using the weighted distance measurement, the topological structure represents the original 3D environmental space of the point cloud and learns the color information simultaneously. In particular, the algorithm of GNGA includes the normalization of each feature vector in the input vector for solving the scalability problem in the 2D image. However, it is difficult to normalize the input vector of the 3D point cloud with the other properties if the point cloud is composed of the increasing environmental map. In addition, the number of clusters generated from the topological structure is only one since it is difficult to cut the edges between each objects from the floor surface (in the green circles of Figure 2c) by designing the appropriate weights from the unknown data distribution composed of the multiple properties’ vector. Therefore, the weighted distance-based methods need a cutting algorithm of the edges for clustering the 2D/3D point cloud data according to the property after generating the topological structure. In this way, the learning method of the 2D/3D point cloud with additional information that can generate the topological structure composed of the multiple properties and preserve the position space simultaneously is not realized.
In this paper, we propose growing neural gas with different topologies (GNG-DT) as a new topological structure learning method for addressing all of these issues, which were not incorporated in the previous work. GNG-DT has multiple topologies of each property, while the conventional GNG method has a single topology of the input vector. In addition, the distance measurement in the winner node selection uses only the position information for preserving the environmental space of the point cloud. The main contributions of this paper are listed as follows:
1.
GNG-DT can preserve 2D/3D position spaces with additional feature information from the point cloud by using the distance measurement of only the position information.
2.
GNG-DT can give multiple clustering results by utilizing multiple topologies within the framework of online learning.
3.
GNG-DT is a robust learning algorithm for scale variance of the input vector composed of point cloud with additional properties.
This paper is organized as follows. Section 2 proposes growing neural gas with different topologies for the 3D space perception. Section 3 shows several experimental results by using 2D and 3D simulation data and RGB-D data. Finally, we summarize this paper and discuss the future direction to realize the 3D space perception.

2. Growing Neural Gas with Different Topologies

2.1. Overview of Algorithm

In this section, we explain the overview of our proposed method called “growing neural gas with different topologies (GNG-DT)”. Figure 3 shows an overall image of the GNG-DT-based point cloud processing method. GNG-DT uses almost the same distance measurement as GNG-U II, and GNG-DT learns the multiple topologies for clustering the point cloud from the different viewpoints of the multiple properties. First, we define the variables used in GNG-DT.
The set S i n of the input vector’s properties is S i n ={Position (pos), Color (col), ...}; the set S r e f of the node’s properties is S r e f ={Position ( p o s ), Color ( c o l ), Normal vector ( n o r ), ...}; and the input vector v and the ith node (reference vector) h are defined as v = ( v p o s , v c o l , ...), h i = ( h i p o s , h i c o l , h i n o r , ...), respectively. Next, a distance measurement of the oth property between the input vector and the ith node is defined as
d i o = v o h i o
In addition, the edge set of the oth property is defined as C o ={ c 1 , 2 o ,…, c i , j o ,…} ( o S r e f ) for generating the multiple topologies in GNG-DT. The detailed algorithm is explained as follows.
  • Step 0. Generate two nodes at random positions, h 1 and h 2 in R n , where n is the dimension of the reference vector. Initialize the connection set c 1 , 2 o = 1 ( o S r e f ), and the age of edge g 1 , 2 is set to 0.
  • Step 1. Generate at random an input data v .
  • Step 2. Select the 1st winner node s 1 and the 2nd winner node s 2 from the set of nodes by
    s 1 = arg min i A d i p o s s 2 = arg min i A s 1 d i p o s
    where A indicates the set of the node numbers.
  • Step 3. Add the squared distance between the input data and the 1st winner to an accumulated error variable E s 1 :
    E s 1 E s 1 + ( d s 1 p o s ) 2
  • Step 4. Set the age of the connection between s 1 and s 2 at 0 ( g s 1 , s 2 = 0). If a connection of the position information between s 1 and s 2 does not yet exist, create the connection ( c s 1 , s 2 p o s = 1 ). In addition, the edge of the other property o( S r e f p o s ) is calculated as the following equation:
    c s 1 , s 2 o = 1 i f h s 1 o h s 2 o < τ o c s 1 , s 2 o = 0 o t h e r w i s e
    where τ o is the predefined threshold value of the oth property.
  • Step 5. Update the nodes of the winner and its direct topological neighbors by the learning rates η 1 and η 2 ( η 1 > η 2 ).
    h s 1 h s 1 + η 1 ( v h s 1 ) h j o h j o + η 2 ( v h j o ) i f c s 1 , j o = 1
  • Step 6. Increment the age of all edges emanating from s 1 .
    g s 1 , j g s 1 , j + 1 i f c s 1 , j p o s = 1
  • Step 7. Delete edges of all properties with an age larger than g m a x . If this results in nodes having no more connecting edges ( c s 1 , s 2 p o s = 0) of the position information, remove those nodes as well.
  • Step 8. If the number of input data generated so far is an integer multiple of a parameter λ , insert a new node as follows.
    • Select the node u with the maximal accumulated error.
      u = arg max i A E i
    • Select the node f with the maximal accumulated error among the neighbors of u.
    • Add a new node r to the network and interpolate its node form u and f.
      h r = 0.5 ( h u + h f )
    • Delete the original edges of all properties between u and f. Next, insert edges of the position property connecting the new node r with nodes u and f ( c u , r p o s = 1, c r , f p o s = 1). The edges of the oth property ( S r e f p o s ) are calculated as the following equation:
      c i , j o = 1 i f h i o h j o < τ o c i , j o = 0 otherwise
    • Decrease the error variables of u and f by a temporal discounting rate α ( 0 α 1 ).
      E u E u α E u E f E f α E f
    • Interpolate the error variable of r from u and f.
      E r = 0.5 ( E u + E f )
  • Step 9. Decrease the error variables of all nodes by a temporal discounting rate β ( 0 β 1 ).
    E i E i β E i ( i A )
  • Step 10. Continue with step 1 if a stopping criterion (e.g., the number of nodes or some performance measure) is not yet fulfilled.
Figure 4 shows the total algorithm of GNG-DT, and bold squares indicate the difference part of the conventional method of the GNG algorithm. In the following sections, we explain the difference points of the conventional algorithm.

2.2. Distance Measurement

In the conventional methods, such as GNGA and GNG-U II, the weighted distance measurement is used in the winner node selection and accumulated error update as the following equation:
s 1 = arg min i A o S w o d i o s 2 = arg min i A s 1 o S w o d i o
E s 1 E s 1 + o S w o ( d s 1 o ) 2
where w o indicates the weight of the oth property. In the weighted distance methods, it is difficult to design the weight for preserving the 3D environmental space appropriately if the input vector is composed of the point cloud and the other feature vectors because the scalability of each feature is different. On the other hand, the distance measurement of GNG-DT uses only the position information of the point cloud shown in Equations (2) and (3) that mean w p o s = 1 and w o = 0 ( o S i n p o s ) in Equations (13) and (14) for learning the accurate position space of the point cloud.

2.3. Clustering from Multiple Topologies

The conventional methods generate only one topology, and the edge is added to the topology if there is no edge between the 1st winner node s 1 and the 2nd winner node s 2 . Therefore, the conventional methods need to use the cutting edge algorithm for clustering the point cloud data after generating the topological structure. On the other hand, GNG-DT generates the multiple topologies of each property including in the property set of the reference vector ( S r e f ). Therefore, GNG-DT needs to add the edges with each feature, except the position information ( c s 1 , s 2 p o s ) for generating the multiple topologies. The edges of the other properties are added by calculating the similarity distance between the nodes of each feature and using the predefined threshold value. Figure 5a–c shows the concept image of the multiple topologies in the point cloud with color information. In this case, GNG-DT can generate three topologies composed of position, color and normal vector.
In the clustering of GNG-DT, the topological structure of each property is utilized by searching the connectivity of each cluster. In addition, GNG-DT can provide additional clustering results by combining the multiple topologies. Figure 5d shows an example of the combination using the topologies of the color and normal vector information ( c i , j c o l · c i , j n o r = 1 ). In this way, the multiple topologies enable to provide the multiple clustering results according to the number of property sets of the reference vector and the combination of each property.

2.4. Learning Rule

Next, we explain the learning rule of GNG-DT. The conventional learning rule is calculated as the following equation:
h s 1 h s 1 + η 1 ( v h s 1 ) h j h j + η 2 ( v h j ) i f c s 1 , j = 1
where c s 1 , j indicates the edge of the conventional GNG method, and η 1 and η 2 indicate learning rates. GNG-DT also uses this learning rule for the 1st winner node s 1 . On the other hand, the topological neighborhood of the 1st winner node is updated if the edge of each property is connected (Equation (5)). In this way, the winner node selection and accumulated error calculation use the distance measurement of only the position information, while the node update uses all of the properties for learning the geometric space and the other feature vectors from the point cloud.

2.5. Feature Extraction from the Topological Structure

While learning the topological structure, the normal vectors are extracted as one of the properties of the reference vector ( h i n o r ) from the reference vector of the position information ( h i p o s ). Figure 6 shows the concept image of the local surface. At first, a local surface (depicted in green circles in Figure 6) of the ith node is composed of the nearest nodes ( c i , j p o s = 1). Next, the weighted center of gravity of the local surface is calculated, and then covariance matrix h ¯ i p o s is calculated as follows.
C o v i = j = 1 k ( h j p o s h ¯ i p o s ) T ( h j p o s h ¯ i p o s ) h ¯ i p o s = h i p o s + j = 1 k ω i , j · h j p o s 1 + j = 1 k ω i , j ω i , j = e x p ( | | h i p o s h j p o s | | 2 μ 2 )
where μ indicates the coefficient. After calculating the covariance matrix, the eigenvectors and values of the matrix are calculated for estimating the normal vector. Then, the normal vector h i n o r is assigned to the eigenvector with the minimum eigenvalue [11]. In this way, the property of the reference vector cloud is calculated by utilizing the topological structure of GNG-DT.

3. Experimental Results

3.1. Simulation Data

3.1.1. Experimental Setup

In this section, we conduct simulation experiments using 2D and 3D point cloud data for verifying the effectiveness of our proposed method. Figure 7 shows the dataset in this experiment. In Figure 7a, the total number of the 2D point cloud composed of the red and green rectangles is 10,000 and the dimensions are (width 300) × (height 300). The centers (x, y) of the two red and green rectangles C1, C2, C3, and C4 are (75, 75), (225, 75), (75, 225), and (225, 225), respectively. In Figure 7b, the total number of the 3D point cloud is 112,500, and the point cloud is composed of a half of a red/brown quadratic prism, a half of a red cylinder, and a brown surface.
The distance measurement of GNG-DT in the winner node selection uses only the position information. Therefore, the scale variance of the position information affects the learning result. In this experiment, the scale factor ( s c l = 10.0, 1.0, 0.1, 0.01) multiplied v p o s of each dataset for verifying the scale variance in GNG-DT. The input vector is composed of v = ( s c l · v p o s , v c o l ) . In this experiment, we compared our proposed method with the conventional GNG and GNG with weighted distance measurement (GNG-W) used in GNGA and GNG-U II. GNG-W does not use the normalization of the point cloud for verifying the robustness of the scale variance. The weights of GNG-W are set to w p o s = 0.8 and w c o l = 0.2 , the parameters used in [47]. In addition, Table 1 shows the parameters used in this experiment. These parameters were also used in [47].

3.1.2. Quantitative Evaluation

In this experiment, we define the following evaluation for verifying the effectiveness of the proposed method.
(1)
Quantization error
The quantization error of each property is defined as follows,
r o = 1 N i = 1 N v i o h s 1 o 2 s 1 = arg min j A ( v i p o s h j p o s )
where s 1 indicates the 1st winner node of the position information. By using Equation (17), we can evaluate the accuracy of the node position and color information included in reference nodes.
(2)
Error of the position information between the center of cluster and true value
Data1 is composed of four rectangles (two red and two green rectangles). The error value between the center of each cluster and true value is defined as follows:
e i = m i p o s c i
where m i p o s indicates the center of gravity of the ith rectangle’s cluster, and c i indicates the true position of the ith rectangle. Equation (18) evaluates the accuracy of the clustering result in each method. In this experiment, the number of trials in each method is set to 50 since the GNG-based methods use random sampling, and we calculate the average and variance value of each evaluation.

3.2. RGB-D Data

3.2.1. Experimental Setup

In this section, we conduct the experiment using 3D point cloud data measured by an RGB-D camera for verifying the effectiveness of the real sensing data. The first column of Figure 12 shows the six datasets used in this experiment. These datasets are measured by Azure Kinect DK and composed of target objects on the floor or desk. In this experiment, we also compared our proposed method with the conventional GNG and GNG-W, and the quantization error is used as the evaluation metric for preserving the 3D position space of the 3D point cloud. The parameters of each method use the same as the above simulation experiment, and the number of trials is 50.

3.2.2. Experimental Result

Table 2 and Table 3 show the results of the quantization error, and Figure 8 and Figure 9 show examples of the learning result. In the quantization error of position information, GNG-DT outperformed the other methods for all of the scales, except s c l = 10.0 in Data2, and GNG-DT preserved the 2D and 3D space structures of each point cloud more accurately. In addition, GNG-DT generated almost the same topological structures of the position ( C p o s ) and color ( C c o l ) properties in the different scales from Figure 8 and Figure 9. The topological structure of color information ( C c o l ) in Figure 8 was separated by the boundary of the different color, and GNG-DT clustered each rectangle in all scales. On the other hand, GNG and GNG-W generated different topological structures in the different scales. In particular, GNG and GNG-W could not preserve the space structure of the point cloud in the cases of s c l = 0.1 and 0.01. This result indicates that the influence of the color information is dominant over the position information. Figure 10 shows that the position information’s quantization error multiplied the inverse of the scale factor s c l . The results of GNG and GNG-W were affected by the scale factor, while GNG-DT did not depend on the scale factor since the error values are almost the same in Figure 10. These results also show that GNG-DT can preserve the space structure of the point cloud without affecting the input vector of the other properties by using only the position information in the distance measurement.
In the results of the quantization error of color information (Table 3), the error values of GNG and GNG-W were lower than GNG-DT in the case of s c l = 1.0 and 10.0, while GNG-DT outperformed the other methods in the other results. When the scale factor s c l is 1.0, the scales of the position and color information are almost the same scale; GNG-W outperformed the other methods. On the other hand, GNG, which can be considered GNG-W whose weights are w p o s = 1.0, w c o l = 1.0, outperformed the other methods in the case of s c l = 10.0 . In this way, the weighted distance approaches need to design the suitable weight according to the scale of each property for learning the point cloud effectively. On the other hand, GNG-DT, which can be considered to be GNG-W, whose weights are w p o s = 1.0, w c o l = 0.0, does not depend on the scale factor, and the results are almost the same value. In addition, the color distribution of the nodes is the same as the point cloud, except the boundary nodes from Figure 8. From these results, GNG-DT can robustly generate the topological structure from the point cloud with color information in the scale differences.
In the result of clustering, the GNG and GNG-W could not cluster four rectangles from the 2D simulation data, while only GNG-DT could cluster four rectangles in all of the scales. Specifically, as the scale factor is 10.0, GNG and GNG-W could not preserve the space structure of the point cloud. This shows that the edges between the same color were not deleted since the color information is a predominant factor in the input vector. On the other hand, the topological structure of GNG-DT ( C c o l ) could cluster the four rectangles in all of the scale factors. Table 4 shows the experimental result of the error e i between the center of cluster and true value in each rectangle in the topological structure C c o l . The results of error had constant values in all of the scales. In Figure 8, the topological structures of color information had isolated nodes in the vicinity of the boundary. As a result of the isolated nodes, the centers of clusters shifted from the true value of each center.
Next, Figure 11 shows an example of the learning result using GNG-DT in Data2. In the topological structure of the position information C p o s (Figure 11a), the number of clusters is 1 since the point cloud of Data2 is composed of two objects on the surface plane. In the topological structure of the color information C c o l (Figure 11b), the number of clusters is 3. The surface point cloud and the part of quadratic prism is clustered as the same cluster since the color is almost the same value. In the topological structure of the normal vector information C n o r (Figure 11c), the number of clusters is 4 according to the surface direction. In the topological structure of the color and normal vector information C c o l · C n o r (Figure 11d), the number of clusters is 5 with segmented red and brown planes in the prism by combining the multiple topological structures of the color and normal vector information. In this way, GNG-DT can generate the multiple clustering results according to each property within the framework of the online learning method, preserving the geometric feature of the point cloud.

3.3. RGB-D Data

3.3.1. Experimental Setup

In this section, we conduct the experiment using 3D point cloud data measured by an RGB-D camera for verifying the effectiveness of the real sensing data. The first column of Figure 12 shows the six datasets used in this experiment. These datasets are measured by Azure Kinect DK and composed of target objects on the floor or desk. In this experiment, we also compare our proposed method with the conventional GNG and GNG-W, and the quantization error is used as the evaluation metric for preserving the 3D position space of the 3D point cloud. The parameters of each method use the same as the above simulation experiment, and the number of trials is 50.

3.3.2. Experimental Result

Table 5 and Table 6 show the experimental results of the quantization error. In Table 5, GNG-DT outperformed the other methods in all quantization errors of the position information. Next, GNG-W outperformed the other methods in most quantization errors of the color information since the weight of the color information has a positive effect on decreasing the error values. However, the GNG-DT’s quantization errors of the color information were smaller than those of GNG-W in (a) and (c). Therefore, the weighted distance based method needs to design suitable weight parameters according to the 3D point cloud measured by an RGB-D camera. On the other hand, GNG-DT preserved the geometric space of the 3D point cloud data accurately, which can be utilized to extract features related to the shape of the point cloud. From the above, the distance using only the position information in GNG-DT is a suitable strategy in the 3D space perception.
Next, Figure 12 shows examples of the learning and clustering results of GNG-DT. GNG-DT could learn the geometric feature and color information simultaneously. In addition, GNG-DT could provide the different clustering results from the point cloud by using topological structures of each property. In particular, in Figure 12, the topological structure of segment 4 comprises surface planes of white and brown boxes by combining the topological structures of color and normal vector information from the real environmental data. In this way, GNG-DT can be applied to the unknown 3D point cloud measured in the real environment and cluster the point cloud from the viewpoint of multiple properties with online learning.

4. Discussion

In the 3D space perception of the autonomous mobile robot, grasping the position between the target object and the robot is an essential capability. Therefore, the accuracy of the position information is the most important for the 3D space perception. In the experiments of the simulation and RGB-D data, GNG-DT performed the other methods, except Data2 (scl = 10.0). This is because the 1st and 2nd winner nodes are selected by using only position information in GNG-DT, which is the important strategy for the 3D space perception. On the other hand, GNG and GNG-W outperformed GNG-DT in the part of the results of the quantization error of the color information. GNG and GNG-W generated some spaces near the boundary of the colors as shown in Figure 8a. On the contrary, GNG-DT generated isolates nodes near the boundary; the color information of these nodes was a mixture of each color information, which affected the quantization error of the color information in the results. However, these isolated nodes represent the boundary of the property, and this information can be utilized for recognizing the shape of the object.
Next, the autonomous mobile robot that performs tasks in the unknown environment needs to recognize the unknown object. For recognizing the unknown objects, the robot needs to extract the invariant of the object by utilizing the multiple properties included in the object. The conventional GNG has only one topological structure, which provides only one clustering result. On the other hand, GNG-DT has topological structures of multiple properties. Therefore, as shown in Figure 11 and Figure 12, GNG-DT can generate multiple clustering results according to the situation and the target object. Furthermore, GNG-DT can generate more clustering results by combining each topological structure. The experimental results of this paper verified the combination of the topological structures by using only the color and normal vector information, but there are many combinations utilizing multiple topological structures. Therefore, we will propose the autonomous cluster generation method according to the situation of the robot by searching suitable combinations from the topological structures.

5. Conclusions

This paper proposed growing neural gas with different topologies (GNG-DT) for perceiving the 3D environmental space from unknown 3D point cloud data. GNG-DT can preserve the geometric feature of the 3D point cloud data by using only the position information in the winner node selection and accumulated error calculation. In addition, GNG-DT has multiple topologies of each property for providing the different clustering results according to the properties within the framework of the online learning. The experimental results showed the effectiveness of the GNG-DT by using simulation and real environmental data measured by an RGB-D camera. The performance of the GNG-DT is much better than that of the other conventional methods. In addition, the clustering results showed that GNG-DT can provide various clustering results by utilizing each topological structure.
In this paper, GNG-DT applied to only static data measured by an RGB-D camera. However, we should verify the dynamic data for realizing 3D space perception since the autonomous robot needs to perceive the 3D environmental space in real time. Therefore, we will propose the learning algorithm for applying dynamic data.

Author Contributions

Conceptualization, Y.T.; methodology, Y.T. and A.W.; software, Y.T. and A.W.; validation, Y.T., A.W. and H.M.; formal analysis, Y.T.; investigation, Y.T.; resources, H.M.; data curation, K.O.; writing—original draft preparation, Y.T.; writing—review and editing, Y.T.; visualization, Y.T.; supervision, T.M. and M.M.; project administration, Y.T.; funding acquisition, Y.T. All authors have read and agreed to the published version of the manuscript.

Funding

JSPS KAKENHI Grant Number 20K19894.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This work was supported by JSPS KAKENHI Grant Number 20K19894.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, H.; Reardon, C.; Parker, L.E. Real-time multiple human perception with color-depth cameras on a mobile robot. IEEE Trans. Cybern. 2013, 43, 1429–1441. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Liu, M. Robotic Online Path Planning on Point Cloud. IEEE Trans. Cybern. 2016, 46, 1217–1228. [Google Scholar] [CrossRef]
  3. Cong, Y.; Tian, D.; Feng, Y.; Fan, B.; Yu, H. Speedup 3-D Texture-Less Object Recognition Against Self-Occlusion for Intelligent Manufacturing. IEEE Trans. Cybern. 2019, 49, 3887–3897. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Shen, L.; Fang, R.; Yao, Y.; Geng, X.; Wu, D. No-Reference Stereoscopic Image Quality Assessment Based on Image Distortion and Stereo Perceptual Information. IEEE Trans. Emerg. Top. Comput. Intell. 2019, 3, 59–72. [Google Scholar] [CrossRef]
  5. Song, E.; Yoon, S.W.; Son, H.; Yu, S. Foot Measurement Using 3D Scanning Model. Int. J. Fuzzy Log. Intell. Syst. 2018, 18, 167–174. [Google Scholar] [CrossRef] [Green Version]
  6. Jeong, I.; Ko, W.; Park, G.; Kim, D.; Yoo, Y.; Kim, J. Task Intelligence of Robots: Neural Model-Based Mechanism of Thought and Online Motion Planning. IEEE Trans. Emerg. Top. Comput. Intell. 2017, 1, 41–50. [Google Scholar] [CrossRef]
  7. Wibowo, S.A.; Kim, S. Three-Dimensional Face Point Cloud Smoothing Based on Modified Anisotropic Diffusion Method. Int. J. Fuzzy Log. Intell. Syst. 2014, 14, 84–90. [Google Scholar] [CrossRef]
  8. Izadi, S.; Kim, D.; Hilliges, O.; Molyneaux, D.; Newcombe, R.; Kohli, P.; Shotton, J.; Hodges, S.; Freeman, D.; Davison, A.; et al. KinectFusion: Real-time 3D reconstruction and interaction using a moving depth camera. In Proceedings of the 24th Annual ACM Symposium on User Interface Software and Technology, Santa Barbara, CA, USA, 16–19 October 2011. [Google Scholar]
  9. Keselman, L.; Woodfill, J.I.; Grunnet-Jepsen, A.; Bhowmik, A. Intel realsense stereoscopic depth cameras. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1–10. [Google Scholar]
  10. Han, X.F.; Jin, J.S.; Wang, M.J.; Jiang, W.; Gao, L.; Xiao, L. A review of algorithms for filtering the 3D point cloud. Signal Process. Image Commun. 2017, 57, 103–112. [Google Scholar] [CrossRef]
  11. Pauly, M.; Gross, M.; Kobbelt, L.P. Efficient simplification of point-sampled surfaces. In Proceedings of the Conference on Visualization’02, Boston, MA, USA, 28–29 October 2002; pp. 163–170. [Google Scholar]
  12. Prakhya, S.M.; Liu, B.; Lin, W. B-SHOT: A binary feature descriptor for fast and efficient keypoint matching on 3D point clouds. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–3 October 2015; pp. 1929–1934. [Google Scholar]
  13. Jin, Y.H.; Lee, W.H. Fast cylinder shape matching using random sample consensus in large scale point cloud. Appl. Sci. 2019, 9, 974. [Google Scholar] [CrossRef] [Green Version]
  14. Guo, Y.; Wang, H.; Hu, Q.; Liu, H.; Liu, L.; Bennamoun, M. Deep Learning for 3D Point Clouds: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 4338–4364. [Google Scholar] [CrossRef]
  15. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  16. Li, Y.; Bu, R.; Sun, M.; Wu, W.; Di, X.; Chen, B. PointCNN: Convolution on x-transformed points. Adv. Neural Inf. Process. Syst. 2018, 31, 820–830. [Google Scholar]
  17. Maturana, D.; Scherer, S. VoxNet: A 3D Convolutional Neural Network for real-time object recognition. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–3 October 2015; pp. 922–928. [Google Scholar]
  18. Kohonen, T. Self-Organizing Maps; Springer: Berlin/Heidelberg, Germany, 2000. [Google Scholar]
  19. Li, J.; Chen, B.M.; Lee, G.H. SO-Net: Self-organizing network for point cloud analysis. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 9397–9406. [Google Scholar]
  20. Martinetz, T.M.; Schulten, K.J. A “neural-gas” network learns topologies. Artif. Neural Netw. 1991, 1, 397–402. [Google Scholar]
  21. Fritzke, B. Unsupervised clustering with growing cell structures. Neural Netw. 1991, 2, 531–536. [Google Scholar]
  22. Fritzke, B. A growing neural gas network learns topologies. Adv. Neural Inf. Process. Syst. 1995, 7, 625–632. [Google Scholar]
  23. Garcia-Rodriguez, J.; Cazorla, M.; Orts-Escolano, S.; Morell, V. Improving 3D Keypoint Detection from Noisy Data Using Growing Neural Gas. In Advances in Computational Intelligence. IWANN 2013. Lecture Notes in Computer Science; Rojas, I., Joya, G., Cabestany, J., Eds.; Springer: Tenerife, Spain, 2013; Volume 7903. [Google Scholar]
  24. Orts-Escolano, S.; García-Rodríguez, J.; Morell, V.; Cazorla, M.; Perez, J.A.S.; Garcia-Garcia, A. 3D Surface Reconstruction of Noisy Point Clouds Using Growing Neural Gas: 3D Object/Scene Reconstruction. Neural Process. Lett. 2016, 43, 401–423. [Google Scholar] [CrossRef] [Green Version]
  25. Satomi, M.; Masuta, H.; Kubota, N. Hierarchical growing neural gas for information structured space. In Proceedings of the 2009 IEEE Workshop on Robotic Intelligence in Informationally Structured Space, Nashville, TN, USA, 30 March–2 April 2009; pp. 54–59. [Google Scholar]
  26. Kubota, N.; Narita, T.; Lee, B.H. 3D Topological Reconstruction based on Hough Transform and Growing NeuralGas for Informationally Structured Space. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010; pp. 3459–3464. [Google Scholar]
  27. Orts-Escolano, S.; Garcia-Rodriguez, J.; Serra-Perez, J.A.; Jimeno-Morenilla, A.; Garcia-Garcia, A.; Morell, V.; Cazorla, M. 3D model reconstruction using neural gas accelerated on GPU. Appl. Soft Comput. 2015, 32, 87–100. [Google Scholar] [CrossRef] [Green Version]
  28. Holdstein, Y.; Fischer, A. Three-dimensional surface reconstruction using meshing growing neural gas (MGNG). Vis. Comput. 2008, 24, 295–302. [Google Scholar] [CrossRef]
  29. Saval-Calvo, M.; Azorin-Lopez, J.; Fuster-Guillo, A.; Garcia-Rodriguez, J.; Orts-Escolano, S.; Garcia-Garcia, A. Evaluation of sampling method effects in 3D non-rigid registration. Neural Comput. Appl. 2017, 28, 953–967. [Google Scholar] [CrossRef] [Green Version]
  30. Saputra, A.A.; Takesue, N.; Wada, K.; Ijspeert, A.J.; Kubota, N. AQuRo: A Cat-like Adaptive Quadruped Robot with Novel Bio-Inspired Capabilities. Front. Robot. 2021, 8, 35. [Google Scholar] [CrossRef]
  31. Saputra, A.A.; Chin, W.H.; Toda, Y.; Takesue, N.; Kubota, N. Dynamic Density Topological Structure Generation for Real-Time Ladder Affordance Detection. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 4–8 November 2019; pp. 3439–3444. [Google Scholar]
  32. Alwaely, B.; Abhayaratne, C. Adaptive Graph Formulation for 3D Shape Representation. In Proceedings of the 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 1947–1951. [Google Scholar]
  33. Rangel, J.C.; Morell, V.; Cazorla, M.; Orts-Escolano, S.; García-Rodríguez, J. Object recognition in noisy RGB-D data using GNG. Pattern Anal. Appl. 2017, 20, 1061–1076. [Google Scholar] [CrossRef] [Green Version]
  34. Rangel, J.C.; Morell, V.; Cazorla, M.; Orts-Escolano, S.; García-Rodríguez, J. Using GNG on 3D Object Recognition in Noisy RGB-D data. In Proceedings of the 2015 International Joint Conference on Neural Networks (IJCNN), Killarney, Ireland, 12–16 July 2015; pp. 1–7. [Google Scholar]
  35. Parisi, G.I.; Weber, C.; Wermter, S. Human Action Recognition with Hierarchical Growing Neural Gas Learning. In Artificial Neural Networks and Machine Learning—ICANN 2014, Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2014; Volume 8681. [Google Scholar]
  36. Mirehi, N.; Tahmasbi, M.; Targhi, A.T. Hand gesture recognition using topological features. Multimed Tools Appl. 2019, 78, 13361–13386. [Google Scholar] [CrossRef]
  37. Yanik, P.M.; Manganelli, J.; Merino, J.; Threatt, A.L.; Brooks, J.O.; Green, K.E.; Walker, I.D. Use of kinect depth data and Growing Neural Gas for gesture based robot control. In Proceedings of the 2012 6th International Conference on Pervasive Computing Technologies for Healthcare (PervasiveHealth) and Workshops, San Diego, CA, USA, 21–24 May 2012; pp. 283–290. [Google Scholar]
  38. do Rego, R.L.M.E.; Araujo, A.F.R.; de Lima Neto, F.B. Growing Self-Organizing Maps for Surface Reconstruction from Unstructured Point Clouds. In Proceedings of the 2007 International Joint Conference on Neural Networks, Orlando, FL, USA, 12–17 August 2007; pp. 1900–1905. [Google Scholar]
  39. Toda, Y.; Matsuno, T.; Minami, M. Multilayer Batch Learning Growing Neural Gas for Learning Multiscale Topologies. J. Adv. Comput. Intell. Intell. Inform. 2021, 25, 1011–1023. [Google Scholar] [CrossRef]
  40. Viejo, D.; Garcia-Rodriguez, J.; Cazorla, M. Combining visual features and growing neural gas networks for robotic 3D SLAM. Inf. Sci. 2014, 276, 174–185. [Google Scholar] [CrossRef] [Green Version]
  41. Orts-Escolano, S.; García-Rodríguez, J.; Cazorla, M.; Morell, V.; Azorin, J.; Saval, M.; Garcia-Garcia, A.; Villena, V. Bioinspired point cloud representation: 3D object tracking. Neural Comput. Applic 2018, 29, 663–672. [Google Scholar] [CrossRef]
  42. Fiser, D.; Faigl, J.; Kulich, M. Growing neural gas efficiently. Neurocomputing 2013, 104, 72–82. [Google Scholar] [CrossRef]
  43. Angelopoulou, A.; García-Rodríguez, J.; Orts-Escolano, S.; Gupta, G.; Psarrou, A. Fast 2D/3D object representation with growing neural gas. Neural Comput. Appl. 2018, 29, 903–919. [Google Scholar] [CrossRef] [Green Version]
  44. Frezza-Buet, H. Online computing of non-stationary distributions velocity fields by an accuracy controlled growing neural gas. Neural Netw. 2014, 60, 203–221. [Google Scholar] [CrossRef]
  45. Angelopoulou, A.; García-Rodríguez, J.; Orts-Escolano, S.; Kapetanios, E.; Liang, X.; Woll, B.; Psarrou, A. Evaluation of different chrominance models in the detection and reconstruction of faces and hands using the growing neural gas network. Pattern Anal. Appl. 2019, 22, 1667–1685. [Google Scholar] [CrossRef] [Green Version]
  46. Molina-Cabello, M.A.; López-Rubio, E.; Luque-Baena, R.M.; Dom’inguez, E.; Thurnhofer-Hemsi, K. Neural controller for PTZ cameras based on nonpanoramic foreground detection. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, Alaska, 14–19 May 2017; pp. 404–411. [Google Scholar]
  47. Tünnermann, J.; Born, C.; Mertsching, B. Saliency From Growing Neural Gas: Learning Pre-Attentional Structures for a Flexible Attention System. IEEE Trans. Image Process. 2019, 28, 5296–5307. [Google Scholar] [CrossRef]
  48. Toda, Y.; Yu, H.; Ju, Z.; Takesue, N.; Wada, K.; Kubota, N. Real-time 3D Point Cloud Segmentation using Growing Neural Gas with Utility. In Proceedings of the 9th International Conference on Human System Interaction, Portsmouth, UK, 6–8 July 2016; pp. 418–422. [Google Scholar]
Figure 1. Examples of topological learning methods. In (e), each color indicates each cluster. (a) Input data, (b) SOM, (c) GCS, (d) GNG, (e) clustering.
Figure 1. Examples of topological learning methods. In (e), each color indicates each cluster. (a) Input data, (b) SOM, (c) GCS, (d) GNG, (e) clustering.
Applsci 12 01705 g001
Figure 2. An example of GNG that applied to simulation data. Circles and lines respectively indicate node and edge of GNG in (b,c). (a) Image data, (b) result of GNG, (c) result of GNG with the weighted vector.
Figure 2. An example of GNG that applied to simulation data. Circles and lines respectively indicate node and edge of GNG in (b,c). (a) Image data, (b) result of GNG, (c) result of GNG with the weighted vector.
Applsci 12 01705 g002
Figure 3. Overview of the major steps in GNG-DT. After measuring 3D point cloud with color from RGB-D camera, the multiple topologies are generated by using competitive learning method. By utilizing and combining the multiple topologies, GNG-DT can provide various clustering results from the viewpoint of each property.
Figure 3. Overview of the major steps in GNG-DT. After measuring 3D point cloud with color from RGB-D camera, the multiple topologies are generated by using competitive learning method. By utilizing and combining the multiple topologies, GNG-DT can provide various clustering results from the viewpoint of each property.
Applsci 12 01705 g003
Figure 4. Flowchart of GNG-DT, where t indicates the number of data input. The boxes drawn by bold line indicate the modified step from the conventional GNG.
Figure 4. Flowchart of GNG-DT, where t indicates the number of data input. The boxes drawn by bold line indicate the modified step from the conventional GNG.
Applsci 12 01705 g004
Figure 5. Examples of multiple topologies of GNG-DT. By combining each topological structure, GNG-DT can provide additional clustering results, such as (d). (a) Position information, (b) color information, (c) normal vector information, (d) color and normal vector information.
Figure 5. Examples of multiple topologies of GNG-DT. By combining each topological structure, GNG-DT can provide additional clustering results, such as (d). (a) Position information, (b) color information, (c) normal vector information, (d) color and normal vector information.
Applsci 12 01705 g005
Figure 6. Concept image of the local surface for extracting normal vector. The surface elements represent green nodes in the dot circle.
Figure 6. Concept image of the local surface for extracting normal vector. The surface elements represent green nodes in the dot circle.
Applsci 12 01705 g006
Figure 7. Experimental dataset. (a) 2D point cloud (Data1), (b) 3D point cloud (Data2).
Figure 7. Experimental dataset. (a) 2D point cloud (Data1), (b) 3D point cloud (Data2).
Applsci 12 01705 g007
Figure 8. Learning examples in simulation experiment using 2D point cloud (Data1). (a) s c l = 1.0 (Left: GNG, Center left: GNG-W, Center right: GNG-DT( C p o s ), Right: GNG-DT( C c o l )), (b) s c l = 10.0 (Left: GNG, Center left: GNG-W, Center right: GNG-DT( C p o s ), Right: GNG-DT( C c o l )), (c) s c l = 0.01 (Left: GNG, Center left: GNG-W, Center right: GNG-DT( C p o s ), Right: GNG-DT( C c o l )).
Figure 8. Learning examples in simulation experiment using 2D point cloud (Data1). (a) s c l = 1.0 (Left: GNG, Center left: GNG-W, Center right: GNG-DT( C p o s ), Right: GNG-DT( C c o l )), (b) s c l = 10.0 (Left: GNG, Center left: GNG-W, Center right: GNG-DT( C p o s ), Right: GNG-DT( C c o l )), (c) s c l = 0.01 (Left: GNG, Center left: GNG-W, Center right: GNG-DT( C p o s ), Right: GNG-DT( C c o l )).
Applsci 12 01705 g008
Figure 9. Learning examples in simulation experiment using 3D point cloud (Data2). (a) s c l = 1.0 (Upper left: GNG, Upper right: GNG-W, Lower left: GNG-DT( C p o s ), Lower right: GNG-DT( C c o l )), (b) s c l = 10.0 (Upper left: GNG, Upper right: GNG-W, Lower left: GNG-DT( C p o s ), Lower right: GNG-DT( C c o l )), (c) s c l = 0.01 (Upper left: GNG, Upper right: GNG-W, Lower left: GNG-DT( C p o s ), Lower right: GNG-DT( C c o l )).
Figure 9. Learning examples in simulation experiment using 3D point cloud (Data2). (a) s c l = 1.0 (Upper left: GNG, Upper right: GNG-W, Lower left: GNG-DT( C p o s ), Lower right: GNG-DT( C c o l )), (b) s c l = 10.0 (Upper left: GNG, Upper right: GNG-W, Lower left: GNG-DT( C p o s ), Lower right: GNG-DT( C c o l )), (c) s c l = 0.01 (Upper left: GNG, Upper right: GNG-W, Lower left: GNG-DT( C p o s ), Lower right: GNG-DT( C c o l )).
Applsci 12 01705 g009
Figure 10. Experimental results of quantization errors ( r p o s ) divided by scale vales ( s c l ). (a) Data1, (b) Data2.
Figure 10. Experimental results of quantization errors ( r p o s ) divided by scale vales ( s c l ). (a) Data1, (b) Data2.
Applsci 12 01705 g010
Figure 11. Clustering examples in 3D point cloud data (Data2). Each color of the node and edge represents each cluster. (a) Position C p o s , (b) color C c o l , (c) normal vector C n o r , (d) color C c o l and normal vector C n o r .
Figure 11. Clustering examples in 3D point cloud data (Data2). Each color of the node and edge represents each cluster. (a) Position C p o s , (b) color C c o l , (c) normal vector C n o r , (d) color C c o l and normal vector C n o r .
Applsci 12 01705 g011
Figure 12. Experimental dataset and results of GNG-DT. The first column is the 3D point cloud measured by an RGB-D camera, the second column is the learning results of GNG-DT, the third column is the clustering results using the topological structure of position information C p o s , the fourth column is the clustering results using the topological structure of color information C c o l , the fifth column is the clustering results using the topological structure of normal vector information C n o r , and the sixth column is the clustering results using the topological structure of color and normal vector information C c o l and C n o r .
Figure 12. Experimental dataset and results of GNG-DT. The first column is the 3D point cloud measured by an RGB-D camera, the second column is the learning results of GNG-DT, the third column is the clustering results using the topological structure of position information C p o s , the fourth column is the clustering results using the topological structure of color information C c o l , the fifth column is the clustering results using the topological structure of normal vector information C n o r , and the sixth column is the clustering results using the topological structure of color and normal vector information C c o l and C n o r .
Applsci 12 01705 g012
Table 1. Setting parameters.
Table 1. Setting parameters.
λ 500
η 1 0.025
η 2 0.0003
g m a x 88
α 0.5
β 0.0005
τ c o l 10.0
τ n o r 0.02
Table 2. Experimental results of quantization error of position in simulation experiment, where GNG, GNG-W and GNG-DT indicate growing neural gas, GNG with weighted distance measurement, and GNG with different topologies, respectively. s c l indicates scale factor.
Table 2. Experimental results of quantization error of position in simulation experiment, where GNG, GNG-W and GNG-DT indicate growing neural gas, GNG with weighted distance measurement, and GNG with different topologies, respectively. s c l indicates scale factor.
Data1GNGGNG-WGNG-DT
s c l = 1.08.687 ± 0.0627.794 ± 0.037 7 . 341 ± 0 . 022
s c l = 10.073.948 ± 0.19173.581 ± 0.217 73 . 434 ± 0 . 199
s c l = 0.11.311 ± 0.0201.144 ± 0.015 0 . 735 ± 0 . 002
s c l = 0.010.248 ± 0.0090.190 ± 0.003 0 . 073 ± 0 . 000
Data2GNGGNG-WGNG-DT
s c l = 1.011.825 ± 0.04711.133 ± 0.038 10 . 341 ± 0 . 021
s c l = 10.0 103 . 064 ± 0 . 311 103.224 ± 0.293103.418±0.322
s c l = 0.11.644 ± 0.0101.456 ± 0.008 1 . 0349 ± 0 . 002
s c l = 0.010.279 ± 0.0030.234 ± 0.002 0 . 103 ± 0 . 000
Table 3. Experimental results of quantization error of color in simulation experiment, where GNG, GNG-W and GNG-DT indicate growing neural gas, GNG with weighted distance measurement, and GNG with different topologies, respectively. s c l indicates scale factor.
Table 3. Experimental results of quantization error of color in simulation experiment, where GNG, GNG-W and GNG-DT indicate growing neural gas, GNG with weighted distance measurement, and GNG with different topologies, respectively. s c l indicates scale factor.
Data1GNGGNG-WGNG-DT
s c l = 1.025.152 ± 0.499 22 . 583 ± 0 . 567 26.178 ± 0.425
s c l = 10.0 21 . 094 ± 0 . 554 23.081 ± 0.62326.168 ± 0.509
s c l = 0.128.619 ± 0.68227.623 ± 0.650 26 . 072 ± 0 . 534
s c l = 0.0133.693 ± 1.89130.425 ± 1.005 26 . 094 ± 0 . 475
Data2GNGGNG-WGNG-DT
s c l = 1.046.740 ± 0.29045.396 ± 0.233 33 . 925 ± 0 . 034
s c l = 10.034.099 ± 0.05533.933 ± 0.028 33 . 919 ± 0 . 032
s c l = 0.148.084 ± 0.49647.841 ± 0.419 33 . 915 ± 0 . 031
s c l = 0.0149.462 ± 1.30148.056 ± 0.766 33 . 915 ± 0 . 034
Table 4. Experimental results of clustering error of GNG-DT, where GNG-DT indicates growing neural gas with different topologies. s c l indicates scale factor.
Table 4. Experimental results of clustering error of GNG-DT, where GNG-DT indicates growing neural gas with different topologies. s c l indicates scale factor.
Cluster No.Cluster Error ( e i )Cluster No.Cluster Error ( e i )
C1 ( s c l = 1.0)5.691 ± 1.520C2 ( s c l = 1.0)5.691 ± 1.809
C1 ( s c l = 10.0)57.654 ± 15.035C2 ( s c l = 10.0)57.562 ± 14.653
C1 ( s c l = 0.1)0.541 ± 0.152C2 ( s c l = 10.0)57.562 ± 14.653
C1 ( s c l = 0.01)0.058 ± 0.019C2 ( s c l = 0.01)0.053 ± 0.014
C3 ( s c l = 1.0)5.518 ± 1.759C4 ( s c l = 1.0)5.830 ± 1.700
C3 ( s c l = 10.0)56.296 ± 19.622C4 ( s c l = 10.0)57.493 ± 16.728
C3 ( s c l = 0.1)0.601 ± 0.191C4 ( s c l = 0.1)0.564 ± 0.165
C3 ( s c l = 0.01)0.055 ± 0.016C4 ( s c l = 0.01)0.056 ± 0.016
Table 5. Experimental results of quantization error of position in real environment, where GNG, GNG-W and GNG-DT indicate growing neural gas, GNG with weighted distance measurement, and GNG with different topologies, respectively.
Table 5. Experimental results of quantization error of position in real environment, where GNG, GNG-W and GNG-DT indicate growing neural gas, GNG with weighted distance measurement, and GNG with different topologies, respectively.
GNGGNG-WGNG-DT
Figure 12a12.796 ± 0.001112.341 ± 0.0018 11 . 976 ± 0 . 0005
Figure 12b12.754 ± 0.19112.914 ± 0.015 12 . 754 ± 0 . 0004
Figure 12c11.428 ± 0.00110.857 ± 0.001 10 . 014 ± 0 . 0005
Figure 12d8.908 ± 0.0908.526 ± 0.088 8 . 309 ± 0 . 109
Figure 12e10.151 ± 0.0169.849 ± 0.023 9 . 725 ± 0 . 043
Figure 12f9.492 ± 0.00069.200 ± 0.0008 8 . 998 ± 0 . 0005
Table 6. Experimental results of quantization error of color in real environment, where GNG, GNG-W and GNG-DT indicate growing neural gas, GNG with weighted distance measurement and GNG with different topologies, respectively.
Table 6. Experimental results of quantization error of color in real environment, where GNG, GNG-W and GNG-DT indicate growing neural gas, GNG with weighted distance measurement and GNG with different topologies, respectively.
GNGGNG-WGNG-DT
Figure 12a33.719 ± 0.71530.544 ± 0.650 30 . 235 ± 0 . 221
Figure 12b14.967 ± 0.076 13 . 876 ± 0 . 077 15.392± 0.047
Figure 12c61.171 ± 0.55557.318 ± 0.343 55 . 773 ± 0 . 271
Figure 12d27.484 ± 0.322 24 . 765 ± 0 . 232 26.320± 0.353
Figure 12e18.925 ± 0.070 17 . 699 ± 0 . 053 18.642± 0.076
Figure 12f24.535 ± 0.355 22 . 426 ± 0 . 229 28.281± 0.268
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Toda, Y.; Wada, A.; Miyase, H.; Ozasa, K.; Matsuno, T.; Minami, M. Growing Neural Gas with Different Topologies for 3D Space Perception. Appl. Sci. 2022, 12, 1705. https://doi.org/10.3390/app12031705

AMA Style

Toda Y, Wada A, Miyase H, Ozasa K, Matsuno T, Minami M. Growing Neural Gas with Different Topologies for 3D Space Perception. Applied Sciences. 2022; 12(3):1705. https://doi.org/10.3390/app12031705

Chicago/Turabian Style

Toda, Yuichiro, Akimasa Wada, Hikari Miyase, Koki Ozasa, Takayuki Matsuno, and Mamoru Minami. 2022. "Growing Neural Gas with Different Topologies for 3D Space Perception" Applied Sciences 12, no. 3: 1705. https://doi.org/10.3390/app12031705

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop