Next Article in Journal
Joint Classification of Hyperspectral and LiDAR Data Based on Adaptive Gating Mechanism and Learnable Transformer
Previous Article in Journal
LinkNet-Spectral-Spatial-Temporal Transformer Based on Few-Shot Learning for Mangrove Loss Detection with Small Dataset
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CUS3D: A New Comprehensive Urban-Scale Semantic-Segmentation 3D Benchmark Dataset

College of Systems and Engineering, National University of Defense Technology, Changsha 437100, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(6), 1079; https://doi.org/10.3390/rs16061079
Submission received: 16 January 2024 / Revised: 8 March 2024 / Accepted: 14 March 2024 / Published: 19 March 2024
(This article belongs to the Section Urban Remote Sensing)

Abstract

:
With the continuous advancement of the construction of smart cities, the availability of large-scale and semantically enriched datasets is essential for enhancing the machine’s ability to understand urban scenes. Mesh data have a distinct advantage over point cloud data for large-scale scenes, as they can provide inherent geometric topology information and consume less memory space. However, existing publicly available large-scale scene mesh datasets are limited in scale and semantic richness and do not cover a wide range of urban semantic information. The development of 3D semantic segmentation algorithms depends on the availability of datasets. Moreover, existing large-scale 3D datasets lack various types of official annotation data, which hinders the widespread applicability of benchmark applications and may cause label errors during data conversion. To address these issues, we present a comprehensive urban-scale semantic segmentation benchmark dataset. It is suitable for various research pursuits on semantic segmentation methodologies. This dataset contains finely annotated point cloud and mesh data types for 3D, as well as high-resolution original 2D images with detailed 2D semantic annotations. It is constructed from a 3D reconstruction of 10,840 UVA aerial images and spans a vast area of approximately 2.85 square kilometers that covers both urban and rural scenes. The dataset is composed of 152,298,756 3D points and 289,404,088 triangles. Each 3D point, triangular mesh, and the original 2D image in the dataset are carefully labeled with one of the ten semantic categories. Six typical 3D semantic segmentation methods were compared on the CUS3D dataset, with KPConv demonstrating the highest overall performance. The mIoU is 59.72%, OA is 89.42%, and mAcc is 97.88%. Furthermore, the experimental results on the impact of color information on semantic segmentation suggest that incorporating both coordinate and color features can enhance the performance of semantic segmentation. The current limitations of the CUS3D dataset, particularly in class imbalance, will be the primary target for future dataset enhancements.

1. Introduction

In the 3D world, the large-scale semantic segmentation of a scene is a task that assigns specific meanings to every object within the 3D scene, aggregating the complex and diverse environmental information into several specific semantic categories to facilitate planning and research. This task helps machines accurately identify object features in large scenes and plays a crucial role in advancing the development of fields such as scene-level robot navigation [1], autonomous driving [2], urban planning [3], spatial analysis [4], and urban fluid simulation [5].
To improve the machine’s ability to recognize the semantics behind urban data, researchers often need ample real data from which the machine can learn. Deep-learning-driven algorithms are currently mainstream in large-scale semantic segmentation methods. With the maturity of convolutional neural networks (CNNs), deep networks that can directly process point clouds (e.g., PointNet [6], PointNet++ [7], KPConv [8], and RandLA-Net [9]) have achieved remarkable results in point-cloud semantic segmentation, especially for city-scale outdoor scenes.
Deep neural networks that directly process mesh data, such as MeshCNN [10] and MeshNet [11], only achieve good results in the semantic segmentation of small-scale objects; they perform poorly in mesh semantic segmentation in large-scale scenes. This phenomenon may be due to the existence of few publicly available large-scale scene mesh datasets with fine-grained semantic annotations. Currently, there is a scarcity of mesh semantic segmentation datasets for outdoor scene, and the ones that do exist have certain limitations. For instance, datasets like ETHZ CVL RueMonge [12] and Hessigheim 3D [13] suffer from multiple labeling errors and unmarked areas. These issues can be attributed to factors like multi-view optimization and indistinct boundaries, leading to label errors in the annotated data. SUM [14] only contains one type of data (mesh) and has only six semantic categories. Thus, it lacks the richness of rural and suburban scene information, making it unsuitable for research on smart urban and rural development planning. Multi-view-based methods for semantic segmentation in large scenes [15] are also among the mainstream approaches for 3D semantic segmentation, but no publicly available large-scale 3D semantic segmentation datasets of scenes with relevant multi-view 2D images currently exist, which somewhat restricts the development of multi-view-based algorithms. Therefore, the current large-scale scene semantic segmentation datasets face the following challenges. First, there is a relatively small number of large-scale scene mesh semantic segmentation datasets, and some of these datasets contain large areas with labeling errors or unlabeled regions. Additionally, the semantic types of scenes are limited and do not include various scenes such as cities, suburbs, and rural areas. Second, these datasets do not provide multiple official data types with precise labels, and errors or omissions in data labeling may occur during the process of the sampling and conversion of 3D data such as point clouds and meshes.
Therefore, we propose a new example of urban-level outdoor semantic segmentation of scenes with rich semantic labels and diverse data types. Our dataset, called CUS3D, includes two types of 3D data: point clouds and meshes, which are annotated with both the real color information and semantic labels. We also provide the original high-resolution aerial images used for 3D reconstruction. Fine-grained semantic annotations are performed on both the 3D data and 2D images, creating a benchmark for current research on 2D and 3D semantic segmentation algorithms. The point-cloud data in this dataset consist of approximately 152,298,756 3D points, while the mesh data consist of 289,404,088 triangles. It covers an area of approximately 2.85 square kilometers and includes 10 semantic categories: building, road, grass, car, high vegetation, playground, water, building site, farmland, and ground. At the same time, we conduct benchmark tests on six classic 3D semantic segmentation deep networks to ensure the suitability of the CUS3D dataset. Compared to existing 3D datasets, CUS3D has the following contributions.
1. CUS3D addresses the problems of the insufficient quantity and high error rate of semantic segmentation datasets for outdoor scene mesh. It provides a new benchmark for large-scale urban scene mesh semantic segmentation. Each triangle face in the dataset has undergone multiple manual cross-checks to ensure the accuracy of the semantic labels.
2. CUS3D is a comprehensive urban-scale outdoor scene semantic segmentation benchmark that also provides official point cloud and high-resolution 2D aerial image data. The semantic labels of 2D images are geometrically consistent with the 3D data, avoiding label errors or omissions in the sampling and data transformation process. It is suitable for testing almost all mainstream 3D semantic segmentation algorithms and can also be used for research in areas such as 3D rendering, 3D reconstruction, and remote sensing image semantic segmentation.
3. CUS3D dataset has richer semantics and covers the semantic information of almost all of the urban scenes. CUS3D also includes the semantic information of suburban and rural scenes, such as farmland and building sites. This benefit provides new opportunities for large-scale urban applications, such as smart cities and urban planning.

2. Related Work

Owing to the diversity of acquisition methods and reconstruction techniques, the data types of 3D data are more varied than those of 2D data. Among them, point clouds, meshes, and voxels are common types of 3D data.

2.1. 3D Point Cloud Dataset

Four categories of existing 3D point cloud datasets exist, based on their scale and scene type: (1) object-level 3D point cloud datasets, (2) indoor scene-level 3D point cloud datasets, (3) outdoor road-level 3D point cloud datasets, and (4) large-scale urban-level 3D point cloud datasets.

2.1.1. Object-Level 3D Point Cloud Datasets

These datasets mainly consist of individual objects of different categories. These include ModelNet [16], ShapeNet [17], ObjectNet3D [18], PartNet [19], ShapePartNet [20], ScanObjectNN [21], and the newly released OmniObject3D [22]. These object-level 3D datasets are commonly used for performance competitions with respect to algorithms such as visual classification and segmentation.

2.1.2. Indoor Scene-Level 3D Point-Cloud Datasets

These types of datasets are usually collected using depth scanners. These datasets include ScanNet [23], SUN RGB-D [24], NYU3D [25], SceneNN [26], and S3DIS [27], released by Stanford University. These datasets are widely used in the early development stage of deep networks for directly processing point clouds. They are suitable for testing 3D semantic segmentation algorithms for small scenes.

2.1.3. Outdoor Road-Level 3D Point-Cloud Datasets

These datasets are usually collected through LiDAR scanners and RGB cameras, including ParisLille-3D [28], SemanticKITTI [29], SemanticPOSS [30], A2D2 [31], Waymo Open Dataset [32], Toronto-3D [33], nuScenes [34], CSPC-Dataset [35], Lyft Dataset [36], Oakland3D [37], Paris-rue-Madame [38], iQmulus [39], and KITTI [40]. Additionally, some datasets simulate road scenes through synthesis to obtain more accurate semantic labels, such as the Synthia Dataset [41]. These datasets are often applied in research on autonomous driving, including target recognition and semantic segmentation.

2.1.4. Large-Scale Urban-Level 3D Point-Cloud Datasets

Large-scale urban-level 3D point-cloud datasets are typically collected in two ways. The first way is through aerial LiDAR scanning, including DALES [42], LASDU [43], DublinCity [44], Semantic3D [45], and ISPRS [46]. The second way is to obtain 2D images through UAV oblique photography and then obtain point-cloud data through 3D reconstruction technology, including Campus3D [47] and SensetUrban [48]. This type of dataset is now more often acquired using the second way. Compared to point-cloud data obtained from aerial LiDAR scanning, this technology offers color information, which can better reflect the structure of urban scenes and is beneficial to smart city planning.

2.2. 3D Mesh Dataset

Mesh is a common representation of 3D data. Currently, relatively few publicly available 3D datasets are based on mesh data. We classify them into two categories: (1) object-level 3D mesh datasets and (2) scene-level 3D mesh datasets.

2.2.1. Object-Level 3D Mesh Datasets

This type of dataset is typically composed of various instances of single objects. In this type of dataset, individual parts of a single object are annotated and segmented, and multiple classes of objects in different poses are categorized. Examples of such datasets include Human Body [49], COSEG [50], SHREC11 [51], MSB [52], and mesh-type 3D data from ModelNet [16]. End-to-end semantic segmentation deep networks for 3D Mesh data are still in their early stage of development. Existing methods such as MeshCNN [10] and MeshNet [11] perform semantic segmentation and classification on mesh data by defining different feature descriptors. The open-source methods validate the reliability of the algorithms using object-level mesh data.

2.2.2. Scene-Level 3D Mesh Datasets

Scene-level 3D mesh datasets are typically obtained through high-quality 3D reconstruction techniques and have geometric topology and high-resolution real scene textures. Currently available indoor 3D mesh datasets include ScanNet [23], Matterport 3D [53], Replica Dataset [54], and 3D-FUTURE [55]. However, relatively few large-scale mesh datasets for outdoor urban-level scenes exist. ETHZ CVL RueMonge [12] is the first benchmark dataset provided in mesh format that is related to urban scenes. However, owing to errors in multi-view optimization and fuzzy boundaries, many incorrect labels exist in this dataset. Hessigheim 3D [13] is a small-scale urban scene semantic segmentation dataset that covers an area of only 0.19 square kilometers. The semantic labels of the mesh data are transferred from the point-cloud data labels, so approximately 40% of the area is unmarked. Some non-manifold vertices also exist, making it difficult to directly apply this dataset. SUM [14] is the latest publicly available outdoor large-scale urban-level 3D mesh dataset known. The dataset covers an area of about 4 square kilometers and includes six semantic categories of urban scenes. However, the SUM [14] dataset only contains mesh data, which are relatively limited in terms of data type and semantic categories. The dataset does not include semantic categories of suburban village scenes for urban and rural planning, so it has certain limitations in practical applications.
In summary, most existing publicly available large-scale 3D datasets consist of a single data type and have limited semantic richness. Especially for large-scale mesh datasets, the number of such datasets is small, and some datasets have partial semantic omissions or marking errors owing to semantic annotation strategies, which affect their use. With the continuous development of smart city planning and other fields, a comprehensive 3D semantic segmentation dataset with multiple data types and rich urban scene semantic information has become meaningful. In the following sections, we introduce a new urban-level outdoor large-scale semantic segmentation comprehensive dataset with diverse data types and rich, accurate semantic annotation. Table 1 compares the Urban3D dataset with other existing 3D urban datasets.

3. Comprehensive Urban-Scale Semantic Segmentation 3D (CUS3D) Dataset

3.1. UAV Captures Aerial Images

In the past, 3D data (such as point clouds) were typically obtained by using LiDAR scanners in cars. With the development of UAV tilt photography and 3D reconstruction technology, UAVs have gained significant advantages in mapping. To obtain high-resolution aerial image sequences, we selected the DJI M300 RTK (https://enterprise.dji.com/cn/matrice-300, accessed on 8 June 2023) (as shown in Figure 1a) and equipped it with the advanced Penta-camera SHARE PSDK 102S V3 (https://shareuav.cn/102PV2, accessed on 12 June 2023) (as shown in Figure 1b) for 2D aerial image acquisition. Figure 1 shows the image acquisition equipment, and Table 2 describes the relevant parameters of the SHARE PSDK 102s camera (https://shareuav.cn/102PV2, accessed on 12 June 2023) (as shown in Figure 1b).
To ensure comprehensive coverage of the designated measurement area, the UAV follows a preplanned Z-shaped flight path and captures images at specified intervals of 2 s. The flight altitude of the UAV is 100 m, and the weather conditions are good without any obstructive factors such as clouds, haze, or strong winds, providing the UAV with good visibility and stable flight conditions. The flight control system is automated. Owing to the limited capacity of the UAV’s onboard batteries, every battery group can support approximately 30 min of flight time. The capture of images for the entire area is composed of multiple separate flight missions executed in parallel. Throughout the data-collection process, the UAV acquired a total of 10,840 aerial images from five distinct angles. Ground Sampling Distance (GSD) is a key metric in photogrammetry and remote sensing for assessing image resolution. In the CUS3D dataset, the original aerial images have a GSD of 2 cm/pixel.
To validate the accuracy and quality of the evaluation data, we use a high-precision, real-time kinematic (RTK) GNSS for geo-referencing the captured aerial images. In addition, six ground control points are manually set up in the real scene to validate the accuracy of the subsequent 3D reconstruction data. The positions of these six checkpoints are the locations where the UAV takes off for data collection. Figure 2 shows the locations of six artificial control points.

3.2. 3D Model Reconstruction

In the reconstruction process of mesh data, the first step is to perform aerial triangulation on the collected 2D image data to determine the position and pose of the camera and minimize the re-projection error. Then, the results of the aerial triangulation are used to construct a large-scale 3D scene using 3D reconstruction techniques. The 3D reconstruction process primarily relies on the Structure From Motion (SFM) theory, which is employed to derive 3D information from a sequence of 2D images capturing motion. The process begins with obtaining the mapping relationship between the pixel of the aerial image and the camera imaging plane through camera calibration. Subsequently, the relative pose transformation relationship between the cameras is calculated by performing feature matching between images. The transformation results from multiple image pairs are refined to accurately determine the camera’s position and orientation, establishing the corresponding relationship between feature point pixels and 3D points. To identify outliers, the RANSAC algorithm is employed to eliminate matching errors of feature points between images, removing outliers to prevent interference with the pose-estimation process. Finally, the sparse point cloud is projected back onto the image as seed points to generate the depth map, establishing the correspondence between all pixels in the image and 3D points.
The correspondence between 2D pixels and 3D points can be described using affine transformations. The process begins by determining the relative pose transformation between cameras through feature point matching in images. An extrinsic matrix is then created to relate the 3D vertex coordinate system to the camera coordinate system. Subsequently, an intrinsic matrix is constructed using camera parameters like focal length to establish the connection between the camera coordinate system and the pixel coordinate system. Ultimately, the relationship between 2D pixels and 3D points is established through coordinate transformations.
However, large lakes exist in our CUS3D dataset, where the 3D water bodies do not meet the Lambertian assumption during reconstruction. To address this issue, we use semi-automated water body restoration techniques to obtain a complete 3D mesh model. The entire area covers an approximate land area of 2.85 square kilometers and consists of approximately 289 million triangular meshes. Figure 3 shows the dimensions and overall appearance of the reconstructed 3D textured mesh model. The terrain in this area is relatively flat, typical of suburban urban scenes, with a diverse range of features including buildings, roads, water systems, farmland, and vegetation. Note that the entire reconstructed scene is divided into 93 tiles, but only 89 tiles contain scene data. There are four tiles without any scene data, and their distribution is shown in Figure 4. These four blank tiles will be ignored in subsequent labeling and experiments. The details of the reconstructed mesh and point-cloud scene are shown in Figure 5.
In the reconstruction process of point-cloud data, we down-sample the mesh data to obtain point-cloud data. We transfer the real color information to the vertices of the point-cloud data through the nearest neighbor search. The generated dense colored point cloud consists of 152,298,756 3D points. The point cloud density is approximately 0.15 m, covering an area of about 2.85 square kilometers.
To verify the accuracy of the 3D reconstruction data, we place six control points manually on the ground while capturing the original images. After completing the construction of the 3D model, we report a median re-projection error of 0.82 pixels and a median reprojection error for every control point of 0.62 pixels, which falls within a reasonable error range. Therefore, we consider the constructed 3D model to be accurate and reliable.

3.3. 3D Semantic Label Annotation

Our CUS3D dataset contains a variety of data types and diverse semantic information. It consists not only of semantic labels for urban scenes but also semantic labels for rural scenes, such as agricultural landscapes and suburban infrastructure. As modern cities grow and smart city initiatives expand, it has become increasingly important to incorporate urban–rural and suburban planning and construction into smart city development. By integrating the semantic information from rural scenes into urban-level 3D models, we can make accurate, informed decisions for smart urban-rural planning.
In the selection of semantic labeling strategies, the CUS3D dataset abandons the previous fixed urban semantic labeling strategy and adopts a dynamic urban semantic labeling strategy that considers both fully developed and developing semantic categories. For example, the categories of “road” and “ground” may have some overlapping characteristics. However, “road” belongs to fully developed functional objects, which can be used for research on urban transportation planning and peak traffic control optimization. “Ground” belongs to undeveloped objects and labeling, and recognition can help with early planning judgments in urban development. “Building sites” belong to the category of ongoing semantic objects and can be transformed into buildings in future semantic updates. The semantic labeling strategy of the CUS3D dataset considers the application, functionality, and temporal aspects of objects, making it more suitable for practical applications such as urban planning, transportation planning, and construction decision-making.
According to the standard definition of semantic categories, we identify 10 meaningful semantic categories in the CUS3D dataset. Considering that the scarcity of certain objects does not affect the planning of and research on large-scale scenes, we categorize some high-granularity object information (e.g., pedestrians, utility poles, and solar panels) into their respective larger categories. These 10 semantic categories comprehensively represent the scene information in cities and suburbs. Every semantic label is assigned a specific color information. Table 3 provides the RGB values and grayscale values corresponding to every semantic label. Every 3D point or triangle mesh in the dataset is assigned a unique semantic label using annotation tools. The specific 10 object classes in our benchmark dataset are as follows:
  • Building, including commercial buildings and residential homes;
  • Road, including streets, asphalt roads, nonmotorized lanes, and parking lots;
  • Car, including small cars and large buses;
  • Grass, including lawns and small shrubs;
  • High vegetation, including fir trees and banyan trees;
  • Playground, including basketball courts, athletic tracks, and amusement park;
  • Water, including lakes, and canals;
  • Farmland, including cultivated fields, undeveloped land, and livestock enclosures;
  • Building sites, including construction material yards, and construction sites;
  • Ground, including cement surfaces, asphalt surfaces, and bare soil surfaces.
Table 3. Semantic label color information table.
Table 3. Semantic label color information table.
CategoryRGB ValueGrayscale Value
Building(254, 1, 252)105
Road(0, 255, 255)179
Car(200, 191, 154)189
Grass(91, 200, 31)99
High vegetation(0, 175, 85)112
Playground(130, 30, 30)0
Water(0, 0, 255)29
Farmland(140, 139, 30)59
Building sites(30, 140, 201)201
Ground(154, 0, 255)75
We use the DP-Modeler software version 2.0 for semi-automated 3D annotation to ensure that every triangle mesh is assigned the corresponding semantic labels. The semantic annotation process starts by inputting the source data, configuring the annotation categories according to requirements, and then performing manual labeling. After labeling has been completed, the classified data are subjected to quality inspection. Abnormal data are reclassified, resulting in labeled output data. The entire annotation process is illustrated in Figure 6, and the tool interface used in the semantic annotation process is shown in Figure 7.
To prevent annotation errors and omissions, we ensure that all of the labels undergo two rounds of manual cross-checking to ensure accuracy. Figure 8 and Figure 9 show the issues that were discovered during the manual cross-checking process and promptly corrected. Figure 10 and Figure 11 show an example of our dataset’s semantic annotation.

3.4. Two-Dimensional Image Data Annotation

In addition to providing 3D meshes and point-cloud data with rich semantic labels, CUS3D also annotates high-resolution raw aerial images captured by a UAV, giving them detailed semantic labels. There is no publicly available 3D semantic segmentation dataset with annotated 2D images. Providing both 2D and 3D data with semantic consistency is beneficial for the development of 3D rendering, 3D reconstruction, and semantic segmentation of remote sensing images. These finely annotated 2D aerial images are also beneficial for training and improving the recognition capability of classic image semantic segmentation networks like DeepLabV3+ [56] on large-scale aerial imagery.
We use an automated instance segmentation tool based on the SAM [57] model for intelligent semantic labeling. This technology is not limited to specific scenarios and can shorten the labeling cycle. Regarding 2D image semantic labeling, the entire 2D image sequence consists of 10,840 images from five different perspectives. Due to the high similarity in the poses of four cameras tilted at 45°, we only selected 4336 image sequences with a 90° top-down view and one 45° oblique view for semantic annotation. We provide 4336 annotated 2D image data, divided into training, testing, and validation sets in an 8:1:1 ratio. The semantic labels of these images are the same as the semantic categories of the 3D models. They have geometric alignment in data annotation, and each image is annotated with high-quality pixel-level annotations and manual cross-checking, ensuring high annotation accuracy. Figure 12 shows the results of our 2D aerial image semantic annotation.

4. Experiment

4.1. Data Split

During the data partitioning for the experiment, a total of 82 tiles were used for the training/validation/testing sets. Among them, fouur tiles did not have mesh data because they were located at the edge of the measurement area, where the number of feature points for 3D reconstruction was too small to construct mesh data. Additionally, seven tiles were not included in the dataset partition because, during the testing of the RandLA-Net network, the point cloud data of these seven tiles had fewer than 4000 points, which did not meet the network’s num_points requirement. In order to ensure the consistency of the input data for the network, these seven tiles were deliberately ignored during the data partitioning.
To use the dataset as a benchmark for semantic segmentation tasks, we divide all of the blocks from the CUS3D dataset in a ratio of 8:1:1. Sixty-six blocks are used as training data, eight blocks as testing data, and eight blocks as validation data (see Figure 13). Owing to the varying sizes of triangles in the 3D mesh data, when evaluating the semantic information of the mesh data, we calculated the surface area occupied by the triangles, rather than the number of triangles. For every one of the ten semantic classes, we calculate the total surface area of the corresponding class in the dataset to represent its distribution, as shown in Figure 13. From the class distribution chart, it can be observed that certain classes, such as the car and playground, account for less than 5% of the total surface area, while categories like building, lake, and high vegetation account for over 70% of the total surface area. Thus, imbalanced class distribution poses a significant challenge in supervised-0learning-based semantic segmentation.

4.2. Baselines

Currently, to our knowledge, no open-source deep learning framework exists that can directly handle large-scale texture mesh data in urban 3D spaces. However, in urban environments, point clouds and meshes have almost identical fixed attributes. Both types of data share the same set of feature vectors in the same scene. Therefore, we choose point-cloud data from the dataset as the input for testing deep neural networks. The test results show consistent accuracy for both types of 3D data. We carefully select six representative methods as 3D benchmark tests for our CUS3D dataset, including supervised and weakly supervised semantic segmentation methods. Table 4 shows the parameter settings of the baseline experiment. In addition, we selected the DeepLabV3+ [56] network to perform semantic segmentation benchmark testing on 2D aerial imagery data.
  • PointNet [6]. This is a groundbreaking work for directly processing unordered point clouds.
  • PointNet++ [7]. This is an improved version of PointNet that includes the extraction of local features.
  • RandLA-Net [9]. This work directly deals with the semantic segmentation of large-scale point clouds and ranked first in the Semantic3D dataset challenge.
  • KPConv [8]. This method references kernel-point convolution for semantic segmentation and achieves good performance on the DALES dataset.
  • SPGraph [58]. This is one of the methods that uses super point graphs to process large-scale point-cloud data.
  • SQN [59]. This is one of the weakly supervised methods that applies a small number of labels for semantic segmentation of large-scale point clouds.
  • DeepLabV3+ [56]. This is the method for benchmark testing for semantic segmentation of 2D aerial images.
Table 4. Baseline experiment parameter settings.
Table 4. Baseline experiment parameter settings.
EpochBatch_SizeNum_PointLearning_RateOptimizerMomentumParametersTime
PointNet1002440960.001Adam0.90.97 M6.4 h
PointNet++2003240960.001Adam0.91.17 M6 h
RandLA-Net1001640960.01Adam0.94.99 M7.5 h
KPConv50010-0.01Adam0.914.08 M8 h
SPGraph500240960.01Adam0.90.21 M5.5 h
SQN1004840960.01Adam0.93.45 M7 h
DeepLabV3+20016-0.007Adam0.910.3 M8 h

4.3. Evaluation Metrics

In the selection of evaluation metrics for semantic segmentation, we chose accuracy, recall, F1 score, and intersection over union (IoU) for every category, similar to the metrics for existing benchmarks. For the overall region testing evaluation, the metrics we adopted are overall accuracy (OA), mean accuracy (mAcc), and mean intersection over union (mIoU). Table 5 shows the baseline test experimental hardware environment configuration.

4.4. Benchmark Results

To ensure fairness in the experiment, we maintained the original experimental environment for every baseline and conducted the experiments on an Nvidia Tesla V100 GPU. The hyper-parameters of every experimental network are adjusted based on the validation data to achieve the best results possible. Table 6 lists the quantitative results of different network experiments after testing. We solely compared the experimental outcomes of the 3D semantic segmentation network. For each class, we chose the best values and emphasized them by highlighting them in bold. The results of the 2D semantic segmentation network are not considered in the comparison. Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 show the test results for every semantic category in the six 3D semantic segmentation networks. Figure 14 and Figure 15 show the visual results of some 2D and 3D baseline experiments.
Our CUS3D dataset provides additional color information to the network compared to the large-scale 3D datasets collected by LiDAR, which is helpful for better understanding the real world. However, this may cause the network to overfit. To explore the impact of RGB information on the final semantic segmentation results, we selected five baseline networks for comparative experiments. In every group of baseline comparative experiments, only coordinate information was input, or both the coordinate information and RGB information were input for training. Table 13 shows the quantitative experimental results of different feature inputs on five baselines.

5. Discussion

On our dataset benchmark, KPConv [8] has the best overall performance in all of the overall performance indicators, with an mIoU of 59.72%, an OA of 89.42%, and a mAcc of 97.88%. The reason KPConv [8] has good overall and individual category performance is that the network has a good ability to learn local features, and it can capture geometric and semantic information in the data. SQN [59] has good overall performance, especially in the few categories with fewer points (e.g., Car, Playground, Building site), thanks to the use of shell-based representations to process point data, which offers good rotation and scaling invariance. At the same time, it can be seen from the results that our dataset does not perform well on SPGraph [58], mainly because it cannot capture local feature information of aggregated data well in large scenes, which also limits the wide application of the CUS3D dataset.
In the experimental results exploring the impact of color information on semantic segmentation, we can see that when coordinate information and color information are input at the same time, PointNet [6], KPConv [8], and RandLA-Net [9] achieve better performance. On the contrary, when only coordinate information is used as input, many urban categories cannot be recognized for segmentation, such as playground, building site, and car categories, which achieve poor segmentation performance. The segmentation performance of SPGraph [58] mainly depends on geometric partitioning, so whether RGB information is available has little effect on its segmentation performance.
The results of the above benchmark test experiment show that our CUS3D dataset can be used as a benchmark for existing 3D semantic segmentation networks, which can help machines better understand 3D scene information. However, the CUS3D dataset currently faces a limitation due to an imbalance in semantic categories within large-scale outdoor scenes. Despite including RGB information, experimental results show significant performance gaps between different categories. For instance, categories such as “Car”, “Playground”, and “Building Site” may not be recognized in some experiments. Primarily, outdoor scene objects are categorized into buildings, roads, grass, and vegetation, with secondary categories like cars and building sites being less represented. This imbalance in semantic categories could potentially impact the evaluation of semantic segmentation methods in the future. One potential research direction to address this data imbalance could involve the utilization of more complex loss functions. We firmly believe that researchers will be able to explore an effective solution to address the issue of class imbalance in large-scale scene datasets in the future.
The CUS3D dataset has significantly contributed to the field of semantic segmentation research by addressing the scarcity issue in current outdoor scene mesh semantic segmentation datasets. Through meticulous manual cross-checks, this dataset ensures the accurate assignment of semantic labels to all scene data, expanding the range of semantic categories available for urban and rural scenes. Moreover, the CUS3D dataset offers official point cloud data to prevent inconsistencies in labels due to data conversion in subsequent research. Additionally, it includes original aerial image sequences with annotated semantic labels for their subsequences, making it a valuable resource for research in remote sensing image semantic segmentation, 3D reconstruction, and 3D rendering.
In the future, real-time photogrammetric reconstruction and semantic segmentation are expected to play crucial roles in the intelligent navigation of unmanned aerial vehicles within urban environments. We expect that the CUS3D dataset, as an accurate and high-resolution 3D semantic segmentation dataset, can promote research in intelligent navigation, intelligent transportation, and intelligent cities. Additionally, to enable the application of the CUS3D dataset in more research fields, we plan to supplement it with hyperspectral image data in the future. Hyperspectral images have high spectral resolution and can provide more detailed information about objects, which is beneficial for the inversion of their physical and chemical properties. This will expand the application of the CUS3D dataset to fields such as ocean remote sensing, vegetation remote sensing, precision agriculture, and atmospheric environmental remote sensing.

6. Conclusions

This paper has introduced an urban-level outdoor large-scale scene dataset with diverse data types and rich semantic labels. The dataset includes two types of 3D data, point-cloud and mesh, covering an accurately annotated area of 2.85 square kilometers. It consists of 10 semantic labels, encompassing the semantic information of both the urban and rural scenes, serving as a benchmark for 3D semantic segmentation. The dataset provides raw 2D images captured by UAVs, accompanied by detailed semantic labels. These images can be used for research in areas such as 3D reconstruction and high-resolution aerial image semantic segmentation. Through extensive benchmark tests, we have verified the applicability of the CUS3D dataset to different semantic segmentation baselines and conducted experiments and discussions on whether RGB information can improve semantic segmentation performance. In future work, we will continue to expand the scene scale and semantic information of CUS3D. We hope that the CUS3D dataset can promote research progress in emerging cyber-physical fields, such as modern smart city construction and digital twins.

Author Contributions

Conceptualization, L.G., Y.L. (Yuxiang Liu). and M.Z.; methodology, L.G. and X.C.; software, L.G.; validation, S.Y.; formal analysis, S.Y.; investigation, L.G.; data curation, X.C.; writing—original draft preparation, L.G. and Y.L. (Yuxiang Liu); writing—review and editing, Y.L. (Yu Liu); visualization, Y.L. (Yuxiang Liu); supervision, Y.L. (Yu Liu); project administration, Y.L. (Yu Liu); funding acquisition, Y.L. (Yu Liu). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China under Grant 62171451.

Data Availability Statement

Our dataset and testing code will be released on https://github.com/CapyLin-G/CUS3D accessed on 31 March 2024 after the paper is accepted.

Acknowledgments

This work has been partially supported by the research project of the National Natural Science Foundation of China on cross-category semi-supervised image semantic segmentation methods.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Valada, A.; Vertens, J.; Dhall, A.; Burgard, W. AdapNet: Adaptive Semantic Segmentation in Adverse Environmental Conditions. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 4644–4651. [Google Scholar]
  2. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The KITTI dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
  3. Chen, R. The Development of 3D City Model and Its Applications in Urban Planning. In Proceedings of the 2011 19th International Conference on Geoinformatics, Shanghai, China, 24–26 June 2011; pp. 1–5. [Google Scholar]
  4. Yaagoubi, R.; Yarmani, M.E.; Kamel, A.; Khemiri, W. HybVOR: A voronoi-based 3D GIS approach for camera surveillance network placement. ISPRS Int. J. Geo-Inf. 2015, 4, 754–782. [Google Scholar] [CrossRef]
  5. García-Sánchez, C.; Philips, D.A.; Gorlé, C. Quantifying inflow uncertainties for CFD simulations of the flow in downtown oklahoma city. Build. Environ. 2014, 78, 118–129. [Google Scholar] [CrossRef]
  6. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  7. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: New York, NY, USA, 2017; Volume 30. [Google Scholar]
  8. Thomas, H.; Qi, C.R.; Deschaud, J.-E.; Marcotegui, B.; Goulette, F.; Guibas, L.J. KPConv: Flexible and Deformable Convolution for Point Clouds. arXiv 2019, arXiv:1904.08889. [Google Scholar]
  9. Hu, Q.; Yang, B.; Xie, L.; Rosa, S.; Guo, Y.; Wang, Z.; Trigoni, N.; Markham, A. RandLA-Net: Efficient Semantic Segmentation of Large-Scale Point Clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11108–11117. [Google Scholar]
  10. Hanocka, R.; Hertz, A.; Fish, N.; Giryes, R.; Fleishman, S.; Cohen-Or, D. MeshCNN: A network with an edge. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef]
  11. Feng, Y.; Feng, Y.; You, H.; Zhao, X.; Gao, Y. MeshNet: Mesh neural network for 3D shape representation. Proc. AAAI Conf. Artif. Intell. 2019, 33, 8279–8286. [Google Scholar] [CrossRef]
  12. Riemenschneider, H.; Bódis-Szomorú, A.; Weissenberg, J.; Van Gool, L. Learning where to classify in multi-view semantic segmentation. In Proceedings of the Computer Vision—ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Lecture Notes in Computer Science. Springer International Publishing: Cham, Switzerland, 2014; Volume 8693, pp. 516–532, ISBN 978-3-319-10601-4. [Google Scholar]
  13. Kölle, M.; Laupheimer, D.; Schmohl, S.; Haala, N.; Rottensteiner, F.; Wegner, J.D.; Ledoux, H. The hessigheim 3D (H3D) benchmark on semantic segmentation of high-resolution 3D point clouds and textured meshes from UAV LiDAR and multi-view-stereo. ISPRS Open J. Photogramm. Remote Sens. 2021, 1, 100001. [Google Scholar] [CrossRef]
  14. Gao, W.; Nan, L.; Boom, B.; Ledoux, H. SUM: A benchmark dataset of semantic urban meshes. ISPRS J. Photogramm. Remote Sens. 2021, 179, 108–120. [Google Scholar] [CrossRef]
  15. Kundu, A.; Yin, X.; Fathi, A.; Ross, D.; Brewington, B.; Funkhouser, T.; Pantofaru, C. Virtual Multi-view Fusion for 3D Semantic Segmentation. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M., Eds.; Springer: Cham, Switzerland, 2020; pp. 518–535. [Google Scholar]
  16. Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3D ShapeNets: A Deep Representation for Volumetric Shapes. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1912–1920. [Google Scholar]
  17. Chang, A.X.; Funkhouser, T.; Guibas, L.; Hanrahan, P.; Huang, Q.; Li, Z.; Savarese, S.; Savva, M.; Song, S.; Su, H.; et al. ShapeNet: An Information-Rich 3D Model Repository. arXiv 2015, arXiv:1512.03012. [Google Scholar]
  18. Xiang, Y.; Kim, W.; Chen, W.; Ji, J.; Choy, C.; Su, H.; Mottaghi, R.; Guibas, L.; Savarese, S. ObjectNet3D: A Large Scale Database for 3D Object Recognition. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 160–176. [Google Scholar]
  19. Mo, K.; Zhu, S.; Chang, A.X.; Yi, L.; Tripathi, S.; Guibas, L.J.; Su, H. PartNet: A Large-Scale Benchmark for Fine-Grained and Hierarchical Part-Level 3D Object Understanding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 909–918. [Google Scholar]
  20. Yi, L.; Kim, V.G.; Ceylan, D.; Shen, I.-C.; Yan, M.; Su, H.; Lu, C.; Huang, Q.; Sheffer, A.; Guibas, L. A scalable active framework for region annotation in 3D shape collections. ACM Trans. Graph. 2016, 35, 1–12. [Google Scholar] [CrossRef]
  21. Uy, M.A.; Pham, Q.-H.; Hua, B.-S.; Nguyen, T.; Yeung, S.-K. Revisiting Point Cloud Classification: A New Benchmark Dataset and Classification Model on Real-World Data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1588–1597. [Google Scholar]
  22. Wu, T.; Zhang, J.; Fu, X.; Wang, Y.; Ren, J.; Pan, L.; Wu, W.; Yang, L.; Wang, J.; Qian, C.; et al. OmniObject3D: Large-Vocabulary 3D Object Dataset for Realistic Perception, Reconstruction and Generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 803–814. [Google Scholar]
  23. Dai, A.; Chang, A.X.; Savva, M.; Halber, M.; Funkhouser, T.; Niessner, M. ScanNet: Richly-Annotated 3D Reconstructions of Indoor Scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5828–5839. [Google Scholar]
  24. Song, S.; Lichtenberg, S.P.; Xiao, J. SUN RGB-D: A RGB-D Scene Understanding Benchmark Suite. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 567–576. [Google Scholar]
  25. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor Segmentation and Support Inference from RGBD Images. In Proceedings of the Computer Vision—ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 746–760. [Google Scholar]
  26. Zhou, B.; Zhao, H.; Puig, X.; Fidler, S.; Barriuso, A.; Torralba, A. Scene Parsing through ADE20K Dataset. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5122–5130. [Google Scholar]
  27. Armeni, I.; Sax, S.; Zamir, A.R.; Savarese, S. Joint 2D-3D-Semantic Data for Indoor Scene Understanding. arXiv 2017, arXiv:1702.01105. [Google Scholar]
  28. Roynard, X.; Deschaud, J.-E.; Goulette, F. Paris-lille-3D: A large and high-quality ground-truth urban point cloud dataset for automatic segmentation and classification. Int. J. Robot. Res. 2018, 37, 545–557. [Google Scholar] [CrossRef]
  29. Behley, J.; Garbade, M.; Milioto, A.; Quenzel, J.; Behnke, S.; Stachniss, C.; Gall, J. SemanticKITTI: A Dataset for Semantic Scene Understanding of LiDAR Sequences. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9297–9307. [Google Scholar]
  30. Pan, Y.; Gao, B.; Mei, J.; Geng, S.; Li, C.; Zhao, H. SemanticPOSS: A Point Cloud Dataset with Large Quantity of Dynamic Instances. In Proceedings of the 2020 IEEE Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA, 19 October–13 November 2020; pp. 687–693. [Google Scholar]
  31. Geyer, J.; Kassahun, Y.; Mahmudi, M.; Ricou, X.; Durgesh, R.; Chung, A.S.; Hauswald, L.; Pham, V.H.; Mühlegg, M.; Dorn, S.; et al. A2D2: Audi Autonomous Driving Dataset. arXiv 2020, arXiv:2004.06320. [Google Scholar]
  32. Sun, P.; Kretzschmar, H.; Dotiwalla, X.; Chouard, A.; Patnaik, V.; Tsui, P.; Guo, J.; Zhou, Y.; Chai, Y.; Caine, B.; et al. Scalability in Perception for Autonomous Driving: Waymo Open Dataset. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 2443–2451. [Google Scholar]
  33. Tan, W.; Qin, N.; Ma, L.; Li, Y.; Du, J.; Cai, G.; Yang, K.; Li, J. Toronto-3D: A Large-Scale Mobile LiDAR Dataset for Semantic Segmentation of Urban Roadways. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 202–203. [Google Scholar]
  34. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuScenes: A Multimodal Dataset for Autonomous Driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11621–11631. [Google Scholar]
  35. Tong, G.; Li, Y.; Chen, D.; Sun, Q.; Cao, W.; Xiang, G. CSPC-dataset: New LiDAR point cloud dataset and benchmark for large-scale scene semantic segmentation. IEEE Access 2020, 8, 87695–87718. [Google Scholar] [CrossRef]
  36. John, H.; Guido, Z.; Luca, B.; Yawei, Y.; Long, C.; Ashesh, J.; Sammy, O.; Vladimir, I.; Peter, O. One thousand and one hours: Self-driving motion prediction dataset. In Proceedings of the 2020 Conference on Robot Learning, Virtual, 16–18 November 2020; Volume 155, pp. 409–418. [Google Scholar]
  37. Munoz, D.; Bagnell, J.A.; Vandapel, N.; Hebert, M. Contextual classification with functional Max-Margin Markov Networks. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 975–982. [Google Scholar]
  38. Serna, A.; Marcotegui, B.; Goulette, F.; Deschaud, J.-E. Paris-rue-Madame database: A 3D mobile laser scanner dataset for benchmarking urban detection, segmentation and classification methods. In Proceedings of the 4th International Conference on Pattern Recognition, Applications and Methods ICPRAM 2014, Angers, France, 6–8 March 2014. [Google Scholar]
  39. Vallet, B.; Brédif, M.; Serna, A.; Marcotegui, B.; Paparoditis, N. TerraMobilita/iQmulus urban point cloud analysis benchmark. Comput. Graph. 2015, 49, 126–133. [Google Scholar] [CrossRef]
  40. Geiger, A.; Lenz, P.; Urtasun, R. Are We Ready for Autonomous Driving? The KITTI Vision Benchmark Suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
  41. Ros, G.; Sellart, L.; Materzynska, J.; Vazquez, D.; Lopez, A.M. The SYNTHIA Dataset: A Large Collection of Synthetic Images for Semantic Segmentation of Urban Scenes. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 3234–3243. [Google Scholar]
  42. Varney, N.; Asari, V.K.; Graehling, Q. DALES: A Large-Scale Aerial LiDAR Data Set for Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 186–187. [Google Scholar]
  43. Ye, Z.; Xu, Y.; Huang, R.; Tong, X.; Li, X.; Liu, X.; Luan, K.; Hoegner, L.; Stilla, U. LASDU: A large-scale aerial LiDAR dataset for semantic labeling in dense urban areas. ISPRS Int. J. Geo-Inf. 2020, 9, 450. [Google Scholar] [CrossRef]
  44. Zolanvari, S.M.I.; Ruano, S.; Rana, A.; Cummins, A.; da Silva, R.E.; Rahbar, M.; Smolic, A. DublinCity: Annotated LiDAR Point Cloud and Its Applications. arXiv 2019, arXiv:1909.03613. [Google Scholar]
  45. McCormac, J.; Handa, A.; Leutenegger, S.; Davison, A.J. SceneNet RGB-D: 5M Photorealistic Images of Synthetic Indoor Trajectories with Ground Truth. arXiv 2017, arXiv:1612.05079. [Google Scholar]
  46. Rottensteiner, F.; Sohn, G.; Jung, J.; Gerke, M.; Baillard, C.; Benitez, S.; Breitkopf, U. The isprs benchmark on urban object classification and 3D building reconstruction. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2012, I–3, 293–298. [Google Scholar] [CrossRef]
  47. Li, X.; Li, C.; Tong, Z.; Lim, A.; Yuan, J.; Wu, Y.; Tang, J.; Huang, R. Campus3D: A Photogrammetry Point Cloud Benchmark for Hierarchical Understanding of Outdoor Scene. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12 October 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 238–246. [Google Scholar]
  48. Hu, Q.; Yang, B.; Khalid, S.; Xiao, W.; Trigoni, N.; Markham, A. Towards Semantic Segmentation of Urban-Scale 3D Point Clouds: A Dataset, Benchmarks and Challenges. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 4975–4985. [Google Scholar]
  49. Maron, H.; Galun, M.; Aigerman, N.; Trope, M.; Dym, N.; Yumer, E.; Kim, V.G.; Lipman, Y. Convolutional neural networks on surfaces via seamless toric covers. ACM Trans. Graph. 2017, 36, 1–10. [Google Scholar] [CrossRef]
  50. The Shape COSEG Dataset. Available online: https://irc.cs.sdu.edu.cn/~yunhai/public_html/ssl/ssd.htm (accessed on 20 December 2023).
  51. Lian, Z.; Godil, A.; Bustos, B.; Daoudi, M.; Hermans, J.; Kawamura, S.; Kurita, Y.; Lavou’e, G.; Nguyen, H.; Ohbuchi, R. SHREC’11 Track: Shape Retrieval on Nonrigid 3D Watertight Meshes. In Eurographics Workshop on 3D Object Retrieval; The Eurographics Association: Eindhoven, The Netherlands, 2011. [Google Scholar]
  52. Shape Analysis Group McGill 3D Shape Benchmark. 2005. Available online: https://www.cim.mcgill.ca/~shape/benchMark/#tools (accessed on 20 December 2023).
  53. Chang, A.; Dai, A.; Funkhouser, T.; Halber, M.; Niebner, M.; Savva, M.; Song, S.; Zeng, A.; Zhang, Y. Matterport3D: Learning from RGB-D Data in Indoor Environments. In Proceedings of the 2017 International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017; pp. 667–676. [Google Scholar]
  54. Straub, J.; Whelan, T.; Ma, L.; Chen, Y.; Wijmans, E.; Green, S.; Engel, J.J.; Mur-Artal, R.; Ren, C.; Verma, S.; et al. The Replica Dataset: A Digital Replica of Indoor Spaces. arXiv 2019, arXiv:1906.05797. [Google Scholar]
  55. Fu, H.; Jia, R.; Gao, L.; Gong, M.; Zhao, B.; Maybank, S.; Tao, D. 3D-FUTURE: 3D furniture shape with TextURE. Int. J. Comput. Vis. 2021, 129, 3313–3337. [Google Scholar] [CrossRef]
  56. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  57. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.-Y.; et al. Segment Anything. arXiv 2023, arXiv:2304.02643. [Google Scholar]
  58. Landrieu, L.; Simonovsky, M. Large-Scale Point Cloud Semantic Segmentation With Superpoint Graphs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4558–4567. [Google Scholar]
  59. Hu, Q.; Yang, B.; Fang, G.; Guo, Y.; Leonardis, A.; Trigoni, N.; Markham, A. SQN: Weakly-Supervised Semantic Segmentation of Large-Scale 3D Point Clouds. In Proceedings of the Computer Vision—ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Springer: Cham, Switzerland, 2022; pp. 600–619. [Google Scholar]
Figure 1. (a) DJI M300 UAV used for data collection; (b) SHARE PSDK 102s five-camera used for data collection.
Figure 1. (a) DJI M300 UAV used for data collection; (b) SHARE PSDK 102s five-camera used for data collection.
Remotesensing 16 01079 g001
Figure 2. When capturing raw aerial images with a UAV, the positions of ground control points are manually set by the personnel on the ground.
Figure 2. When capturing raw aerial images with a UAV, the positions of ground control points are manually set by the personnel on the ground.
Remotesensing 16 01079 g002
Figure 3. Original unmarked 3D texture mesh model. High-resolution texture mesh model reconstructed in 3D from a sequence of aerial images taken by a UAV, covering an area of 2.85 square kilometers.
Figure 3. Original unmarked 3D texture mesh model. High-resolution texture mesh model reconstructed in 3D from a sequence of aerial images taken by a UAV, covering an area of 2.85 square kilometers.
Remotesensing 16 01079 g003
Figure 4. Blank Tile Distribution Chart (as indicated by the red box in the image) and Ignored Tile Distribution Chart (as indicated by the blue box in the image).
Figure 4. Blank Tile Distribution Chart (as indicated by the red box in the image) and Ignored Tile Distribution Chart (as indicated by the blue box in the image).
Remotesensing 16 01079 g004
Figure 5. Three-dimensional reconstruction data display: (a) 3D reconstructed textured mesh; (b) texture-less mesh with white base model; (c) 3D point cloud reconstructed by sampling.
Figure 5. Three-dimensional reconstruction data display: (a) 3D reconstructed textured mesh; (b) texture-less mesh with white base model; (c) 3D point cloud reconstructed by sampling.
Remotesensing 16 01079 g005
Figure 6. The pipeline of semantic tagging work. (a) Input the data; (b) Configure semantic label categories; (c) Assign semantic labels to the data; (d) Manually cross-check the labels for correctness; (e) Reassign labels for any incorrect ones; (f) Complete the data annotation.
Figure 6. The pipeline of semantic tagging work. (a) Input the data; (b) Configure semantic label categories; (c) Assign semantic labels to the data; (d) Manually cross-check the labels for correctness; (e) Reassign labels for any incorrect ones; (f) Complete the data annotation.
Remotesensing 16 01079 g006
Figure 7. Semantic labeling software work interface.
Figure 7. Semantic labeling software work interface.
Remotesensing 16 01079 g007
Figure 8. In the manual cross-check, find and adjust the incorrectly labeled wall. (a) Discover the wall that is incorrectly labeled. (b) Select the object with the lasso tool. (c) Assign the correct label to the incorrectly labeled wall.
Figure 8. In the manual cross-check, find and adjust the incorrectly labeled wall. (a) Discover the wall that is incorrectly labeled. (b) Select the object with the lasso tool. (c) Assign the correct label to the incorrectly labeled wall.
Remotesensing 16 01079 g008
Figure 9. Artificial cross-checking detection of unmarked vehicles. (a) Detect unmarked vehicles; (b) select unmarked vehicles; (c) change the marking to vehicles.
Figure 9. Artificial cross-checking detection of unmarked vehicles. (a) Detect unmarked vehicles; (b) select unmarked vehicles; (c) change the marking to vehicles.
Remotesensing 16 01079 g009
Figure 10. Partial region semantic labeling results. (a) Partial region original mesh. (b) Partial region semantic labeling results.
Figure 10. Partial region semantic labeling results. (a) Partial region original mesh. (b) Partial region semantic labeling results.
Remotesensing 16 01079 g010
Figure 11. CUS3D point-cloud benchmark dataset example.
Figure 11. CUS3D point-cloud benchmark dataset example.
Remotesensing 16 01079 g011
Figure 12. Partial high-resolution aerial images and their detailed semantic labels, including 10 semantic categories.
Figure 12. Partial high-resolution aerial images and their detailed semantic labels, including 10 semantic categories.
Remotesensing 16 01079 g012
Figure 13. Baseline test data partition schematic diagram. (a) The distribution of the training, testing, and validation dataset. (b) Semantic category distribution in the training, validation, and testing datasets.
Figure 13. Baseline test data partition schematic diagram. (a) The distribution of the training, testing, and validation dataset. (b) Semantic category distribution in the training, validation, and testing datasets.
Remotesensing 16 01079 g013
Figure 14. Qualitative results of PointNet++ [7], KPConv [8], RandLA-Net [9], and SQN [59] on the test set of CUS3D dataset.
Figure 14. Qualitative results of PointNet++ [7], KPConv [8], RandLA-Net [9], and SQN [59] on the test set of CUS3D dataset.
Remotesensing 16 01079 g014
Figure 15. The results of testing the 2D aerial images in the CUS3D dataset with DeepLabV3+ [56].
Figure 15. The results of testing the 2D aerial images in the CUS3D dataset with DeepLabV3+ [56].
Remotesensing 16 01079 g015
Table 1. Comparison of existing 3D urban benchmark datasets.
Table 1. Comparison of existing 3D urban benchmark datasets.
NameYearData TypeSpatial SizeClassesPoints/TrianglesRGBPlatforms
ISPRS [24]2012Point cloud-91.2 MNoALS
DublinCity [31]2019Point cloud 2.0 km213260 MNoALS
DALES [29]2020Point cloud 10.0 km28505.3 MNoALS
LASDU [30]2020Point cloud 1.02 km253.12 MNoALS
Campus3D [33]2020Point cloud 1.58 km224937.1 MYesUAV camera
SensetUrban [34]2021Point cloud 6.0 km2132847.1 MYesUAV camera
Oakland3D [23]2009Point cloud1.5 km51.6 MNoMLS
Paris-rue-Madame [25]2014Point cloud0.16 km1720 MNoMLS
iQmulus [26]2015Point cloud10.0 km8300 MNoMLS
Semantic3D [32]2017Point cloud-84000 MNoTLS
Paris-Lille-3D [14]2018Point cloud1.94 km9143 MNoMLS
SemanticKITTI [15]2019Point cloud39.2 km254549 MNoMLS
Toronto-3D [19]2020Point cloud1.0 km878.3 MYesMLS
ETHZ RueMonge [44]2014Mesh0.7 km91.8 MYesAuto-mobile camera
Hessigheim 3D [45]2021Point cloud and Mesh 0.19 km211125.7 M/36.76 MYesLidar and camera
SUM [46]2021Mesh4 km2619 MYesUAV camera
CUS3D (Ours)2023Point cloud and Mesh 2.85 km210152.3 M/289.4 MYesUAV camera
Table 2. The parameters of the SHARE PSDK 102s camera.
Table 2. The parameters of the SHARE PSDK 102s camera.
Performance ParameterNumerical Value
Lens number5
Tilt angle 45
Image resolution6144 × 4096
Focal lensDownward: 25 mm, sideward: 35 mm
Sensor sizeAPS-C Format (23.5 × 15.6 mm)
Table 5. Baseline test experimental hardware environment configuration table.
Table 5. Baseline test experimental hardware environment configuration table.
NameModel
System4029GP-TRT2
CPUIntel Xeon 4210R
MemorySAMSUNG 32 GB DDR4 ECC 293
System DiskIntel S4510
Data DiskIntel S4510
GPUNvidia Tesla V100
Table 6. Benchmark results of the baselines on CUS3D. Overall Accuracy (OA, %), mean class accuracy (mAcc, %), mean IoU (mIoU, %), and per-class IoU (%) are reported.
Table 6. Benchmark results of the baselines on CUS3D. Overall Accuracy (OA, %), mean class accuracy (mAcc, %), mean IoU (mIoU, %), and per-class IoU (%) are reported.
GroundBuildingRoadHigh VegetationLakeCarPlay-
Ground
Farm-
Land
GrassBuilding SitemIouOAmAcc
PointNet78.3665.8178.3473.3225.2220.358.768.0239.1116.6642.5474.5794.91
Point-
Net++
25.7880.3657.7786.1857.5236.1311.3840.8840.1713.2855.1684.5296.90
RandLA-Net15.2976.1849.5278.4262.3049.8935.9142.1139.7429.8652.9877.1295.42
KPConv25.7382.8948.6989.6448.4344.6925.7241.8644.8415.5059.7289.4297.88
SPGraph12.0532.1030.5663.4031.714.47.6536.6631.3628.8430.4467.2340.73
SQN30.6578.6360.2279.0859.2155.4634.7739.8845.5430.7356.8278.1995.64
DeepLab- V3+81.2180.2582.4792.5565.1259.5638.6846.7445.1840.0263.1890.0298.04
The best performance of 3D baseline tests for each semantic category is indicated by bold and underline.
Table 7. Evaluation metric results for PointNet for every category.
Table 7. Evaluation metric results for PointNet for every category.
GroundBuildingRoadHigh VegetationWaterCarPlaygroundFarmlandGrassBuilding Site
Acc (%)95.9696.4790.4498.0392.1498.7799.9099.9086.7299.18
Recall (%)88.3580.9591.7788.8040.4225.329.989.9850.5320.10
F1 Score (%)87.8579.3487.8484.5640.1733.5816.1716.1756.2128.40
IoU (%)78.3665.8178.3473.3225.2220.358.768.7639.1116.66
Table 8. Evaluation metric results for PointNet++ for every category.
Table 8. Evaluation metric results for PointNet++ for every category.
GroundBuildingRoadHigh VegetationWaterCarPlaygroundFarmlandGrassBuilding Site
Acc (%)90.4393.6296.1193.6699.1699.1399.8697.5695.1697.57
Recall (%)42.3789.6283.1491.7265.0146.8318.6472.4253.4216.09
F1 Score (%)40.7789.0773.0792.5572.4552.6118.6757.5056.8922.58
IoU (%)25.7880.3657.7786.1857.5236.1311.3840.8840.1713.28
Table 9. Evaluation metric results for RandLA-Net for every category.
Table 9. Evaluation metric results for RandLA-Net for every category.
GroundBuildingRoadHigh VegetationWaterCarPlaygroundFarmlandGrassBuilding Site
Acc (%)88.2494.4995.1290.1598.6797.6497.3296.6793.3196.59
Recall (%)42.5680.5283.0280.4481.3355.3157.3149.5670.6865.00
F1 Score (%)26.0486.3865.1787.8876.4366.0451.8758.8456.4245.38
IoU (%)15.2976.1849.5178.4262.3049.8935.9142.1139.749.86
Table 10. Evaluation metric results for KPConv for every category.
Table 10. Evaluation metric results for KPConv for every category.
GroundBuildingRoadHigh VegetationWaterCarPlaygroundFarmlandGrassBuilding Site
Acc (%)93.7995.0097.5093.9499.5199.2699.9098.2196.1898.76
Recall (%)35.4392.3880.8095.7160.4158.8713.8462.3954.8317.83
F1 Score (%)40.1990.5465.1694.5263.8160.2712.5857.8761.5325.14
IoU (%)25.7382.8948.6989.6448.4344.6925.7241.8644.8415.50
Table 11. Evaluation metric results for SPGraph for every category.
Table 11. Evaluation metric results for SPGraph for every category.
GroundBuildingRoadHigh VegetationWaterCarPlaygroundFarmlandGrassBuilding Site
Acc (%)92.7369.5996.0321.9099.0398.9899.8997.1692.3998.09
Recall (%)14.2438.4842.6890.8840.514.758.9151.3042.1442.56
F1 Score (%)21.3048.4846.5977.4947.918.7816.0453.3047.5446.18
IoU (%)12.0532.1030.5663.4031.714.407.6536.6631.3628.84
Table 12. Evaluation metric results for SQN for every category.
Table 12. Evaluation metric results for SQN for every category.
GroundBuildingRoadHigh VegetationWaterCarPlaygroundFarmlandGrassBuilding Site
Acc (%)86.0995.3695.0691.9898.2898.0497.8796.5293.2296.45
Recall (%)40.2690.1272.0795.9271.0062.7473.3147.4764.6165.87
F1 Score (%)46.4487.7674.7487.7373.8170.7050.9456.3761.9646.32
IoU (%)30.6578.6360.2279.0759.2155.4634.7739.8845.5430.73
Table 13. Quantitative results for the five selected baselines on the CUS3D dataset. Overall Accuracy (OA, %), mean class Accuracy (mAcc, %), mean IoU (mIoU, %), and per-class IoU (%) are reported.
Table 13. Quantitative results for the five selected baselines on the CUS3D dataset. Overall Accuracy (OA, %), mean class Accuracy (mAcc, %), mean IoU (mIoU, %), and per-class IoU (%) are reported.
OA (%)mAcc (%)mIou (%)GroundBuildingRoadHigh
Vegetation
LakeCarPlay-
ground
Farm-
land
GrassBuilding Site
PointNet
(w/o RGB)
54.1790.8320.4338.6723.4759.2022.5812.170.7203.1521.852.18
PointNet
(w/ RGB)
74.5794.9142.5478.3665.8178.3473.3225.2220.358.768.0239.1116.66
PointNet++
(w/o RGB)
70.0394.029.6211.3550.6433.7365.3215.4915.13016.7821.290.14
PointNet++
(w/ RGB)
84.5296.9055.1625.7880.3657.7786.1857.5236.1311.3840.8840.1713.28
RandLA-Net
(w/o RGB)
70.1994.0440.6610.1573.7630.6970.9844.3942.878.0636.4424.5218.32
RandLA-Net
(w/ RGB)
77.1295.4252.9815.2976.1849.5278.4262.3049.8935.9142.1139.7429.86
KPConv
(w/o RGB)
85.0596.2744.5217.4677.1237.5584.4125.6834.0712.5624.5330.430
KPConv
(w/ RGB)
89.4297.8859.7225.7382.8948.6989.6448.4344.6925.7241.8644.8415.50
SQN
(w/o RGB)
69.3193.8641.2524.4770.5048.9367.1244.2440.053.7628.4726.5715.32
SQN
(w/ RGB)
78.1995.6456.8230.6578.6360.2279.0859.2155.4634.7739.8845.5430.73
SPGraph
(w/o RGB)
67.2339.4030.1212.5430.4928.6163.6230.854.297.0437.4630.7229.17
SPGraph
(w/ RGB)
67.2340.7330.4412.0532.1030.5663.4031.714.47.6536.6631.3628.84
The best performance for each semantic category is indicated by bold and underline.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gao, L.; Liu, Y.; Chen, X.; Liu, Y.; Yan, S.; Zhang, M. CUS3D: A New Comprehensive Urban-Scale Semantic-Segmentation 3D Benchmark Dataset. Remote Sens. 2024, 16, 1079. https://doi.org/10.3390/rs16061079

AMA Style

Gao L, Liu Y, Chen X, Liu Y, Yan S, Zhang M. CUS3D: A New Comprehensive Urban-Scale Semantic-Segmentation 3D Benchmark Dataset. Remote Sensing. 2024; 16(6):1079. https://doi.org/10.3390/rs16061079

Chicago/Turabian Style

Gao, Lin, Yu Liu, Xi Chen, Yuxiang Liu, Shen Yan, and Maojun Zhang. 2024. "CUS3D: A New Comprehensive Urban-Scale Semantic-Segmentation 3D Benchmark Dataset" Remote Sensing 16, no. 6: 1079. https://doi.org/10.3390/rs16061079

APA Style

Gao, L., Liu, Y., Chen, X., Liu, Y., Yan, S., & Zhang, M. (2024). CUS3D: A New Comprehensive Urban-Scale Semantic-Segmentation 3D Benchmark Dataset. Remote Sensing, 16(6), 1079. https://doi.org/10.3390/rs16061079

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop