Next Article in Journal
Deep Ground Filtering of Large-Scale ALS Point Clouds via Iterative Sequential Ground Prediction
Previous Article in Journal
Coupling of SWAT and DSAS Models for Assessment of Retrospective and Prospective Transformations of River Deltaic Estuaries
Previous Article in Special Issue
Towards Classification of Architectural Styles of Chinese Traditional Settlements Using Deep Learning: A Dataset, a New Framework, and Its Interpretability
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semantic Segmentation for Digital Archives of Borobudur Reliefs Based on Soft-Edge Enhanced Deep Learning

1
Research Organization of Science and Technology, Ritsumeikan University, Shiga 525-8577, Japan
2
College of Information Science and Engineering, Ritsumeikan University, Shiga 525-8577, Japan
3
Nara National Research Institute for Cultural Properties, Nara 630-8577, Japan
4
Research Center for Area Studies, National Research and Innovation Agency, Jakarta 12710, Indonesia
5
Borobudur Conservation Office, Magelang 56553, Indonesia
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(4), 956; https://doi.org/10.3390/rs15040956
Submission received: 11 December 2022 / Revised: 4 February 2023 / Accepted: 7 February 2023 / Published: 9 February 2023
(This article belongs to the Special Issue Digitization and Visualization in Cultural Heritage)

Abstract

:
Segmentation and visualization of three-dimensional digital cultural heritage are important analytical tools for the intuitive understanding of content. In this paper, we propose a semantic segmentation and visualization framework that automatically classifies carved items (people, buildings, plants, etc.) in cultural heritage reliefs. We also apply our method to the bas-reliefs of Borobudur Temple, a UNESCO World Heritage Site in Indonesia. The difficulty in relief segmentation lies in the fact that the boundaries of each carved item are formed by indistinct soft edges, i.e., edges with low curvature. This unfavorable relief feature leads the conventional methods to fail to extract soft edges, whether they are three-dimensional methods classifying a three-dimensional scanned point cloud or two-dimensional methods classifying pixels in a drawn image. To solve this problem, we propose a deep-learning-based soft edge enhanced network to extract the semantic labels of each carved item from multichannel images that are projected from the three-dimensional point clouds of the reliefs. The soft edges in the reliefs can be clearly extracted using our novel opacity-based edge highlighting method. By mapping the extracted semantic labels into three-dimensional points of the relief data, the proposed method provides comprehensive three-dimensional semantic segmentation results of the Borobudur reliefs.

Graphical Abstract

1. Introduction

Recently, with the development of three-dimensional (3D) digitization of large-scale cultural heritage properties [1,2,3,4], efficient analysis of digitized properties has been the focus of an increasing amount of research [5,6,7]. Segmentation and visualization of three-dimensional scanned cultural heritage are important analytical tools for interpretation and intuitive understanding of the associated content. Semantic segmentation for three-dimensional point clouds has been widely investigated, such as outdoor scene understanding for autonomous driving [8,9,10,11] and robotics [12,13,14]. Semantic segmentation of three-dimensional digital cultural heritage has broad applications, such as the identification of architectural elements [15] or analysis of the state of conservation of materials [16]. However, automatic semantic segmentation for three-dimensional point clouds of cultural heritage is challenging because most cultural heritage properties consist of complex structures and are distinct from each other. With the development and application of deep learning in recent years, the semantic segmentation of three-dimensional point clouds in cultural heritage has made a significant breakthrough [15,16,17].
In this study, we focus on semantic segmentation for three-dimensional point cloud-type digital archives of cultural heritage reliefs. Cultural reliefs were often carved to illustrate stories about historical characters or cultural events. For example, the Borobudur temple, which is a UNESCO Cultural Heritage Site in Indonesia, has 2672 bas-relief panels (sculptural reliefs in which forms extend only slightly from the background), containing 1460 narrative panels and 1212 decorative panels. Figure 1 shows a photograph of the Borobudur Temple in Indonesia. Figure 2 shows an example of the Borobudur Temple relief. These reliefs are the largest collection in the world and can be divided into five sections, each of which describes a different Buddhist story. Semantic segmentation of the reliefs can help researchers understand the figures and objects that are relevant to the stories and distinguish them from the decorative background.
However, semantic segmentation for bas-reliefs has inherent difficulties that are not present in conventional deep learning-based semantic segmentation methods, which focus on indoor or outdoor scene analysis. (1) Relief is an undulated surface different from typical three-dimensional objects that have 360-degree information. This leads to the inability of existing deep learning-based three-dimensional point cloud semantic segmentation methods to work directly on reliefs. Because supervised learning methods require a large annotated dataset, the lack of a benchmark dataset of bas-reliefs makes the task more challenging. (2) The boundaries of each carved item are formed by indistinct soft edges, i.e., edges with low curvature. This unfavorable feature of reliefs leads the conventional methods to fail to extract soft edges, whether they are three-dimensional methods classifying a three-dimensional scanned point cloud or two-dimensional (2D) methods classifying pixels in a drawn image. However, the soft edges in reliefs serve as an important cue in carved item semantic segmentation. An effective edge highlighting method would improve the segmentation accuracy.
In this paper, we propose a semantic segmentation method that automatically classifies carved items (people, buildings, plants, etc.) in cultural heritage reliefs. This method is successfully applied to the bas-reliefs of the Borobudur Temple. To address difficulty (1), we propose a deep learning-based semantic segmentation framework for relief point clouds by projecting them into multichannel images, i.e., RGB images, depth maps, and soft-edge images. In addition, we propose an interactive relief panel extraction method to improve the accuracy of image projection, as well as the efficiency of processing large-scale point clouds during dataset preparation. For difficulty (2), we propose a method that automatically extracts the soft edges as one of the inputs in multichannel images so that the soft edges can be processed specifically. We adopt a novel three-dimensional edge detector referred to as opacity-based edge highlighting. This method is an application of the opacity-controlling mechanism employed in stochastic point-based rendering (SPBR), a high-quality transparent visualization method. The opacity-based edge highlighting method can express the soft-edge area in a relief as a gradation of surface opacity. This opacity gradation produces distinct brightness gradation in the two-dimensional image space, creating recognizable soft-edge images. We utilize the extracted soft edge as guide information in our network. An intermediate prediction of the semantic edge is arranged for the soft-edge information in the network. By supervised multitask learning, soft-edge information is effectively utilized for semantic segmentation.

2. Related Work

As a crucial tool regarded for analysis, segmentation and classification of digital cultural heritage have become widely investigated. Grilli et al. [16] proposed a texture-based approach and a geometry-based approach for the classification of three-dimensional data of heritage scenarios, which are effective for restoration and documentation purposes. Several works have attempted to classify cultural heritage images by employing different kinds of techniques, such as the naive Bayes nearest neighbor classifier [18], support vector machines [19], K-means algorithms [20], and latent Dirichlet allocation [21]. Deep learning has enabled a breakthrough in this work [15,22]. Pierdicca [15] improved the dynamic graph convolutional neural network (DGCNN) to segment elements, such as churches, chapels, cloisters, porticoes, and loggias in architectural cultural heritage three-dimensional point cloud data.
The three-dimensional semantic segmentation aims at classifying each point as a specific part of the point cloud. PointNet [23], an end-to-end deep neural network, is a pioneering and well known point cloud segmentation architecture that operates directly on point clouds and allows for the acquisition of input permutation invariance. PointNet++ [24] was proposed as a development from PointNet to retain local geometries and to consider local neighborhoods to ensure strong robustness and detail capture. PointNet and PointNet++ are representatives of point-based approaches [25,26,27,28,29,30]. In addition, benefiting from the convolutional neural network (CNN), projection-based [31,32,33], voxel-based [34,35], and multirepresentation fusion-based [36,37] methods have been proposed.
As relief is an undulated surface without 360-degree information, we consider that the semantic segmentation of the relief point cloud is suitable to be solved by the projection-based method with two-dimensional semantic segmentation. Two-dimensional semantic segmentation has been studied so successfully in computer vision [38,39,40,41,42,43,44,45,46,47] that increasing research on three-dimensional semantic segmentation is being performed by two-dimensional methods to achieve better results, e.g., projection-based methods [32]. CNNs have been widely used in two-dimensional semantic segmentation [38,39,40,41,42,43,44,45]. Existing segmentation architectures typically consist of two stages: backbone and segmentation. The first stage extracts features from RGB images, with popular models such as ResNet [38] and DenseNet [39] pretrained on the ImageNet dataset [40]. The final stage seeks to make predictions based on the extracted features. In this stage, methods such as upsampling [41], PPM [42], and ASPP [39] are used. Note that the convolutional layers serve as the primary building blocks in both stages. CNN-based semantic segmentation in the overall framework is commonly used, such as FCN [44], SegNet [41], U-Net [45], PSPNet [42], and DeepLab [43].
We utilize a three-dimensional edge highlighting method to extract soft edges as unique features from the relief point cloud. Three-dimensional edge highlighting is more generally analyzed as the feature highlighting of three-dimensional point clouds, which are actively performed by a variety of mathematical and computational technique-based methods [48]. Recently, statistical methods using eigenvalue-based three-dimensional feature values have been widely employed [49,50,51,52]. These three-dimensional feature values are defined based on the eigenvalues of the local three-dimensional covariance matrix. We utilize an opacity-based edge highlighting method [53] to express the soft-edge area in relief as a gradation of surface opacity to confirm that the comparably higher-curvature region can also be recognized. This method is based on SPBR [54,55] to extract soft edges in relief data. SPBR is a transparent visualization method proposed in our previous work. The SPBR algorithm achieves controlled perspective visualization by interpreting opacity as the probability that each pixel in the image is the color of the input point.

3. Method

This section describes the framework of semantic segmentation of the Borobudur reliefs. We introduce the overview of the entire framework in Section 3.1. Our proposed framework contains four parts using different proposed methods. The detailed introductions of these methods are described from Section 3.2, Section 3.3, Section 3.4 and Section 3.5.

3.1. Overview

We propose a single-view projection-based semantic segmentation framework for the three-dimensional point cloud of Borobudur reliefs. The entire structure of the framework is depicted in Figure 3. The proposed method projects information from a three-dimensional point cloud to multichannel images and utilizes a deep learning-based network to predict the semantic segmentation results. First, a relief panel extraction method is proposed to extract the relief, which is the component for semantic segmentation, from photogrammetric point cloud data. This relief panel extraction method, which automatically includes RGB images and depth-map projections, improves the efficiency and accuracy of dataset creation for semantic segmentation. Then, the opacity-based edge highlighting method is utilized to extract the soft-edge information, which is the unique feature of the Borobudur reliefs from the point cloud. After the RGB images, depth maps, and soft-edge images are prepared, we implement a semantic segmentation network based on CNN deep learning named the soft-edge enhanced network. The prepared three kinds of images are concatenated into multichannel images and used as the input of the proposed network. The use of multichannel images improves the accuracy of the proposed semantic segmentation network from different perspectives. First, RGB images are used to project color information from point clouds and convert photogrammetric point clouds into photographs. Second, the depth map addresses the loss of depth information due to the projection. Third, soft-edge images can reflect the slight change in curvature in the relief point clouds. With the proposed method, the corresponding semantic labels of each carved item in reliefs can be extracted. Finally, by mapping the extracted semantic labels into three-dimensional points of the relief data, the proposed method provides comprehensive three-dimensional semantic segmentation results for the Borobudur reliefs.

3.2. Relief Panel Extraction and RGB-D Image Projection

The three-dimensional point cloud of the Borobudur reliefs is obtained by photogrammetry. In the raw data of the scanned points, regions other than relief panels are also included. Therefore, a relief panel extraction method is proposed in this paper. Open-source software, such as CloudCompare [56] and Meshlab [57], can be used to manually extract relief panels. However, manual extraction and adjustment require considerable labor, and the accuracy of image projection cannot be guaranteed. In this work, we propose a semiautomatic method for relief panel extraction and RGB-D image projection. Figure 4 shows the process of relief panel extraction.
Control point selection. The position, size, and direction of the targeted relief panel are determined by selecting four control points. The four points are the four corners of the relief panel and are interactively selected by the user in the clockwise direction starting from the top left corner. Here, the selected points are represented as p 1 , p 2 , p 3 , and p 4 .
Approximate region of interest extraction. This process removes surrounding points and extracts an approximate region of interest around the target relief panel from the raw point cloud data. The average coordinates, p c , of the four control points are used as the center of the spherical region of interest, while the farthest distance from the centroid among the four control points is utilized as the radius r s of the spherical region. Here, p c and r s are calculated using the following equations:
p c = p 1 + p 2 + p 3 + p 4 4
r s = MAX p i p c ,   p i p 1 , p 2 , p 3 , p 4 .
Determination of the relief panel’s local coordinates. Calculating the local coordinates of the relief panel is necessary for correct image projection. This step focuses on calculating the normal vector n of the region of interest by plane fitting. After the second step, most undesired points are trimmed, leaving only the data containing the entire point cloud of the relief panel and some surrounding parts. The normal of the fitting plane is used to set the z-axis of the relief panel’s local coordinate system. The positive direction of the normal is determined by the left-hand rule based on clockwise rotation of the four control points. The vertical direction from the center point p c to the line between the first and second control points is set to be the y-axis of the local coordinate system. The orthogonal x-axis is then set accordingly. After the local coordinate system is determined, the global coordinates of the point cloud can be transformed to local coordinates using the coordinate transformation
C local = R   T   C global
where C global are the global coordinates of the points in the relief panel’s point cloud, and C local are the local coordinates. T is the translation matrix, which translates the origin of the local coordinate system to the origin of the global coordinate system. R is the rotation matrix, which transforms the direction between the two coordinate systems. We utilize a gradually decreasing rectangle to the region of interest to update the local coordinate system automatically.
Relief panel extraction and image projection. After the optimized local coordinate system is determined, the projection of the four control points on the x-y plane forms a quadrilateral. This quadrilateral defines the final region of interest of the target relief panel. Finally, the extracted relief panel is projected to a RGB image and a depth map. The intensity in the depth map is calculated from the value of the z-axis by a linear transformation. The intensity is set in the range of 0 to 255.

3.3. Opacity-Based Edge Highlighting

This section describes our proposed soft-edge extraction method for Borobudur reliefs. Figure 5 shows the process of soft-edge extraction from relief photogrammetric point cloud. We proposed this opacity-based edge highlighting method in our previous work [54]. The main idea of this method is that we extract the three-dimensional edge areas and then execute rendering with higher opacity to the three-dimensional edges to make them brighter than the surrounding non-edge regions.
The opacity is based on SPBR [54]. In this rendering method, the transparency of a visualized surface is achieved based on probabilistic theory. That is, we define the surface opacity by the probability that each pixel becomes the point color [55]. SPBR realizes fast and precise three-dimensional see-through imaging transparent visualization. As the transparency originates from the stochastic determination of pixel intensities, it achieves the correct depth feel in visualization without requiring time-consuming depth sorting of three-dimensional points. Thus, this method is particularly suitable for large-scale three-dimensional point clouds.
The steps to execute SPBR are as follows:
STEP 1. Creation of point ensembles: the point cloud is randomly divided into multiple subgroups, each of which has the same number of three-dimensional points. We call the subgroups “point ensembles.” The number of ensembles is denoted as L.
STEP 2. Point projection per point ensemble: for each point ensemble in step 2, its constituent three-dimensional points are independently projected into the image plane. As a result, L intermediate images are created. In the projection process, we incorporate the point–occlusion effect per pixel.
STEP 3. Averaging the intermediate images: an average image is created by averaging the intensities of the corresponding pixels in the above L intermediate images. Then, the transparent image of the point ensembles can be created. The ensemble number L controls the statistical accuracy and works as an image-quality parameter.
Considering a local area with an area of S, we execute the point–density adjustment for this area, such that the following user-defined surface opacity α is realized:
α = 1 1 s S n adj / L
where s is the particle cross-section area, which is tuned such that an image of each three-dimensional point overlaps exactly one pixel; n adj is the adjusted number of three-dimensional points in this local area; and L is the number of point ensembles.
Based on this opacity provided by SPBR, we can execute our proposed opacity-based edge highlighting method. To extract the three-dimensional edge, the eigenvalue-based three-dimensional feature is utilized to extract high-curvature areas from the given point cloud. Several kinds of eigenvalue-based three-dimensional feature values have been proposed. In our research, we adopt the change in curvature and the linearity:
Change   of   curvature :       C λ = λ 3 λ 1 + λ 1 + λ 3
Linearity :       L λ = λ 1 λ 2 λ 1
where λ 1 , λ 2 , and λ 3 are the eigenvalues of the three-dimensional structure tensor (the local three-dimensional covariance matrix) [58] with λ 1 λ 2 λ 3 0 . In our implementation, the three-dimensional structure tensor is calculated for each spherical local region centered at each point. The change in curvature C λ measures the minimal extension of the local point distribution that should vanish in the case of a planar distribution. The linearity, L λ , measures the difference between the two independent-directional largest extensions of the local point distribution that should also vanish in the case of a planar distribution. Thus, a large value of C λ or L λ indicates that the examined local set of points forms a three-dimensional edge of the point-based surface.
We denote the adopted three-dimensional feature value as f and normalize it to make the distribution fall between zero and one. With the three-dimensional feature value f , we propose a function to relate a three-dimensional feature value to opacity α. The function can highlight the soft edge from the relief point cloud. In this function, we define the opacity α as follows:
α f = α min α max     α min F th     f th d f     f th d   +   α min α max       f f th f th < f F th F th < f 1
where f t h and F t h are two threshold values with f th F th . The parameter d controls the speed of the opacity increase. The opacity graph is shown in Figure 6. In the soft-edge regions, we need a plateau area at f > F th to cover all of the highest curvature regions. Thus, we assign a constant high opacity for local areas with f > F th . To make the opacity gradually decrease as the position moves away from the centerline of the edge regions, we make α gradually increase for local areas with f th < f F th . This opacity distribution gives the narrow areas around the centerlines comparably higher opacity, and the areas are highlighted with brighter colors.
The visualization of soft-edge highlighting from relief photogrammetry by the opacity-based method is shown in Figure 5a. The soft edge is a unique feature in the reliefs that forms boundaries between the carved items and the wall background. Soft edges have only slightly higher curvatures than their surrounding areas. This method successfully extracts the soft-edge features in the relief.
After the soft-edge component is extracted by the proposed method, we project it into two-dimensional images as the input for the network. The value of each pixel in the projected image is converted from the opacity of the extracted three-dimensional edge points, as shown in Equation (8).
g α = 1 α α min α max α min
where g represents the value of each pixel in the projected soft-edge image, and α represents the opacity of the extracted soft-edge points. The soft-edge projection image is shown in Figure 5b.

3.4. Soft-Edge Enhanced Network for Semantic Segmentation

In this section, we present our proposed soft-edge enhanced semantic segmentation network. The proposed method is based on an encoding–decoding structure, as shown in Figure 7. Section 3.4.1 presents the soft-edge enhanced utilization techniques of the semantic segmentation network. The structure of our soft-edge enhanced network is described in Section 3.4.2. The training loss function is illustrated in Section 3.4.3.

3.4.1. Soft-Edge Enhanced Utilization

As the soft edge is a unique feature of relief, we utilize two techniques to effectively apply the soft-edge information. (1) Feature skip connection. This technique is inspired by U-Net [45], a pioneering method in biomedical image segmentation, which is similar to relief analysis, with fewer training data and focusing on high-frequency information, such as edges. During deep convolution, the network tends to focus more on wide perception and ignore the edge detail information. The feature skip connection can recover and retrace edge detail features to enhance the soft-edge information. (2) Multitask learning to supervise the intermediate layer. This technique is inspired by DecoupleSegNets [59], which is based on the concepts of semantic bodies and edges. We utilize a similar idea that decouples the body and the edge with different types of supervision by multitasking to optimize the edge feature and the body feature. The edge feature is supervised by the segmentation edge mask, while the body feature is supervised by the mask where the segmentation edges are ignored during training. Compared with DecoupleSegNets, we focus on edge prediction, while DecoupleSegNets is oriented more to body prediction. Our use of soft-edge information extracted from the three-dimensional edge highlighting as input can facilitate network prediction of the segmentation edges from soft-edge information. Supervising the intermediate layer of the segmentation edge creates an easier task for soft edge information to maintain the clarity of the edge for segmentation and the consistency of body parts for each object.

3.4.2. Network Structure

The network is a multichannel input network with RGB images, depth maps, and soft-edge images. The input layers of the network consist of three different images with a resolution of H, as shown in Figure 7. The RGB image input layer has three channels, and the depth-map input layer and soft-edge input layer have one channel.
At the beginning of the encoder phase, these three input images are concatenated into one integrated feature map with five channels. Then, a subsequent 1 × 1 convolutional layer is utilized to output a feature map that self-adaptively adjusts the number of channels to three channels. The following parts of the encoder contain five layers that are extracted from DenseNet121 [39], with H/2, H/4, H/8, H/16, and H/32. The channel numbers of each layer are 64, 64, 128, 256, and 512.
Three decoder stages that have similar decoder blocks are utilized for the semantic body, edge, and final output. The decoder block first passes the input features through the 3 × 3 convolutional layer twice with a size of H/32. This is followed by four upsampling steps. After each upsampling layer, the feature layer is concatenated with the incoming feature layer of the same size by skip connection, and the features are passed through the 3 × 3 convolution layer twice. The output of the decoder block is the same size as H/2 with 64 channels.
The segmentation edge decoder passes the features of the output of the decoder block into a 3 × 3 convolution layer that outputs with the size H/2 with 1 channel. The sigmoid activation layer is applied to predict the probability of the pixels being inside the segmentation edge region. The segmentation edge consists of the nonzero pixels obtained from performing high-pass filtering on the semantic segmentation result.
In the segmentation body decoder, the output features of the decoder block are decoupled by subtracting the output of the segmentation edge decoder block. A 3 × 3 convolution layer follows the decoupled feature layer that outputs with the size H/2 with N cls channels, where N cls represents the number of classes. Then, a softmax layer is used to determine the probability of each pixel belonging to each class in the segmentation body. The segmentation body consists of the pixels where the segmentation edges are ignored. The learning objective F body for the segmentation body decoder is given by:
F body = F gt F edge
where F gt represents the ground truth of the semantic segmentation, and F edge represents the segmentation edge.
In the final decoder, every two intermediate feature maps extracted by 3 × 3 convolution layers in the segmentation edge and body encoder are involved in the convolution in this decoder block by a skip connection. There is also a 3 × 3 convolution layer and a SoftMax layer following. The 3 × 3 convolution layer outputs the size H/2 with N cls channels, and the softmax layer determines the probability of each pixel belonging to each class in the semantic segmentation result.

3.4.3. Training Loss

To improve the final segmentation accuracy, three loss functions are utilized for the segmentation edge decoder, segmentation body decoder, and final decoder in our paper. These three loss functions are presented as L edge , L body , and L final . Then, a joint loss function L is defined in Equation (10) to train the proposed network. The total training loss L is defined as follows:
L = α L edge s edge ,   s edge + β L body s body ,   s body + γ L final s ,   s
where s edge , s body , and s represent the ground truth of the segmentation edge, the segmentation body, and the final result, respectively; s edge , s body , and s represent the intermediate prediction results from the semantic edge, the segmentation body, and the final decoders, respectively; and α , β , and γ are three hyperparameters that control the weighting among the three losses. We set α = 1.5 , β = 1.0 , and γ = 1.0 as default.
In the segmentation edge decoder, the MSE loss is utilized to predict the probability that a pixel point is inside the segmentation edge region. The expression of L edge is as follows:
L edge s edge ,   s edge = s edge s edge 2
We utilize the focal loss [60] for the probability of segmentation map prediction in the segmentation body and the final decoders. The L body can be defined by Equation (12):
L body s body ,   s body = FL s body M edge ,   s body M edge
L final s ,   s = FL s ,   s
where M edge represents the ground truth of the segmentation edge as a mask. The focal loss is an extension of the cross-entropy loss function that downweights easy examples and focuses training on hard negatives.
FL p t = α 1 p t γ log p t
where p t represents the probability for each class, and where α and γ are two hyperparameters; α is a weighting factor that controls the balance among each class, and γ smoothly adjusts the rate at which easy examples are downweighted.

3.5. Mapping and Visualization

We achieve the three-dimensional semantic segmentation results by mapping the two-dimensional predicted results of the network into the three-dimensional point cloud of the relief panels that are extracted from Section 3.2. The two-dimensional results overlap with the plane of z = 0 in the local coordinate system of the relief. The three-dimensional segmentation results in global coordinates in the photogrammetric point cloud data are obtained from the following equation.
C global = T 1   R 1   C local ,
where C global , C local , T, and R appear in Equation (3) in Section 3.2.
The alpha blending method is utilized to visualize the three-dimensional semantic segmentation result following Equation (16).
c vis = α c cls + 1 α c rgb ,
where α represents an opacity from 0 to 1 and where. c vis , c cls , and c rgb represent the colors of the visualization, segmentation, and original point, respectively.

4. Experimental Results

In this section, we introduce the process of building the dataset for semantic segmentation in Section 4.1. Section 4.2 describes our experimental approach and experimental results for the proposed network. The visualization results of the semantic segmentation for the Borobudur reliefs are shown in Section 4.3.

4.1. Dataset Creation

The training dataset is composed of pairs of RGB images, depth maps, soft-edge images, and segmentation labels. The dataset creation process is shown in Figure 8. The original data are photogrammetric point cloud data that contain the reliefs and structures of the temple. We collect the relief parts that contain three-dimensional coordinates and color information and separate the information into an RGB image and the corresponding depth map by the method introduced in Section 3.2. Moreover, we apply the edge highlighting method introduced in Section 3.3 to extract the soft-edge points and project them into images. We draw the semantic segmentation category labels on the two-dimensional image and set the category to four classes: background (black), characters (red), plants (greed), and others (blue). We annotate 26 images by manually drawing object contours and giving labels to each object.
The resolution of the projected images is 3200 × 1024, which incurs a considerable computational cost to the network. There are only 26 pairs of these projected images, which are not enough to train a deep neural network. Therefore, we cut these images into 6656 patch pairs to train the neural network. The resolution of the image patch is set by three different sizes: 1024 × 1024, 768 × 768, and 512 × 512. Among 26 pairs of projected images, 24 pairs are applied to train the neural network, and two pairs serve as validation data for the quantitative evaluation.

4.2. Network Evaluation

To evaluate the proposed method, two relief panels were chosen for the quantitative and qualitative experiments. From left to right in Figure 9 are the RGB image, depth map, soft edge, ground truth of semantic segmentation, and prediction results. The soft edges feature a difference in concentration. The extraction of soft edges is currently affected by the point cloud density, requiring manual fine-tuning of parameters. However, the experimental results show that our network is robust for different concentrations in soft edges and maintains the clarity of the segmentation edge and the consistency of body parts for each object.
We employed five metrics (accuracy, precision, recall, intersection over union (IoU), and F1-score) and output the confusion matrix to evaluate our network. The results are presented in Table 1. The left side of the table is the confusion matrix, where each row represents the pixel count of each category in the ground truth, and each column represents the pixel count of each category in the predicted results. The right side of the table shows the evaluation metrics on each category; higher values are better. According to the evaluation metrics, our network achieves 90.53% accuracy in prediction, as well as a 81.52% mean IoU.
We also implemented prediction for the segmentation edge information and output the image of the intermediate layer of the segmented edge. The results are shown in Figure 10. Regarding the prediction of edge information, the first column of the figure shows the segmentation edge obtained from the semantic segmentation ground truth, and the second column shows the probability of segmentation edge pixel prediction. The third column shows the ground truth, and the fourth column shows the prediction results of semantic segmentation. We consider that learning the segmentation edges helps the network clarify the contours of semantic segmentation.
We compare our proposed network with the existing two-dimensional semantic segmentation networks SegNet, U-Net, PSPNet, and DeepLabV3+.
The results in Figure 11 show that our proposed network prediction maintains the clarity of the segmentation edge and the consistency of semantic segmentation for each object. A comparison of the evaluation metric results is shown in Table 2.

4.3. Visualization Results

We used a voting algorithm for the divided patches to predict the semantic segmentation results for the whole panel of reliefs. The method mentioned in Section 3.5 used to map the visualization results on the region of interest relativity coordinate.
We performed prediction experiments for semantic segmentation of the whole relief’s photogrammetric point cloud data. The point cloud of the relief is shown at the top of Figure 12. We transform the point cloud into a 3200 × 1024 image using the method described in Section 3. After the point cloud is imaged, we crop the image into N equally spaced and overlapping square 1024 × 1024 image patches. N is usually set to 64. The semantic segmentation results of these N patches are then predicted. Voting is performed on the overlapping parts. The two-dimensional prediction results are shown in the middle part of Figure 12. The final mapping result is shown at the bottom of the figure. Figure 13 shows the visualization results of some additional samples of relief point clouds. Our results achieve very clear classification and high accuracy and improve the readability of the relief content. We also conducted experiments to apply our proposed semantic segmentation framework directly to the original point cloud. The visualization results are shown in Figure 14.

5. Discussion

According to the prediction results in Figure 11, our results outperform those of the conventional two-dimensional semantic segmentation network. Our method is robust to relief images with mottled colors due to corrosion. The bottom 4 rows in Figure 11 show that the colors of the input RGB images are uniform, as these reliefs are not heavily corroded and the segmentation can be visually distinguished. The existing two-dimensional semantic segmentation networks are more stable for these images. However, the top 2 rows in Figure 11 show images of heavily corrupted reliefs whose colors are nonuniform. Their results in the existing two-dimensional semantic segmentation networks are strongly affected, and it is difficult to segment accurately from these unclear colors. This is proven by the results in Table 2, where the existing networks have a low average in recall, precision, accuracy, and mIoU.
However, our results predict consistent semantic segmentation for each object in all cases and are robust to any kind of RGB image. Moreover, our semantic segmentation results make the contours much clearer. A comparison of the results of the evaluation metrics also shows that our network outperforms the other networks by more than 20 percent in all evaluations.
For the proposed method, the resolution of the output image is 3200 × 1024. When mapped to the point cloud, one pixel corresponds so less than 1 mm. The resolution of the segmentation results can successfully support our requirements for segmenting four categories of target objects. Expansion of segmentation categories is a plan for the continuation of this research. There are many other important categories, such as the category “ship” in Borobudur reliefs. This requires high costs of data acquisition. We plan to expand the segmentation categories to enrich semantic segmentation results in our future work.

6. Conclusions

In this paper, we propose a deep-learning-based, soft-edge enhanced semantic segmentation method that automatically classifies carved items (people, buildings, plants, etc.) in cultural heritage reliefs. This method is successfully applied to the bas-reliefs of the Borobudur Temple. The method is characterized by utilizing the unique feature of reliefs, which are the soft (low-curvature) edges composing boundaries of the carved items. The soft edge is extracted by our novel opacity-based edge highlighting method. We also propose a soft-edge enhanced network with a multichannel image as input, i.e., RGB images, depth maps, and soft-edge images. The soft-edge images are utilized as guide information in our soft-edge enhanced network. An intermediate prediction of the semantic edge is arranged for the soft edge information in the network. By supervised multitask learning, soft-edge information is effectively utilized for semantic segmentation. We demonstrate the accuracy of the proposed method. According to the results of the quantitative experiments, the accuracy of the semantic segmentation reaches 90%, which prominently exceeds the accuracy of other current methods. For the results of qualitative experiments, the proposed method successfully classifies people, plants, backgrounds, and other objects and maintains the clarity of the edge for segmentation and the consistency of semantic segmentation for each object. Each kind of pattern is clearly divided in the semantic results, which increases the readability of the contents of the reliefs.
In future work, we consider that, not only the semantic segmentation task, but also the object detection and instance segmentation algorithms can go further to increase the readability of relief data. In our future research, we plan to combine semantic segmentation, target detection, and instance segmentation algorithms to improve our relief visualization and analysis tool. We also hope to use our relief visualization analysis tool for digital museums, humanities research, and other applications.

Author Contributions

Conceptualization, S.J., J.P., L.L., K.H. and S.T.; methodology, S.J., J.P., K.H., L.L. and S.T.; software, S.J.; validation, S.J.; formal analysis, S.J.; investigation, S.J.; resources, H.Y., F.I.T., U.S. and B.; data curation, S.J., J.P. and L.L.; writing—original draft preparation, S.J.; writing—review and editing, S.J., J.P., L.L., K.H. and S.T.; visualization, S.J.; supervision, J.P., L.L. and S.T.; project administration, S.T.; funding acquisition, S.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work is partially supported by JSPS KAKENHI Grant Number 19KK0256, 21H04903, and the Program for Asia-Japan Research Development (Ritsumeikan University, Kyoto, Japan).

Data Availability Statement

Not applicable.

Acknowledgments

In this paper, images of the Borobudur temple are presented with the permission of the Borobudur Conservation Office and Research Center for Area Studies, National Research and Innovation Agency, Indonesia. Additionally, photogrammetry data were acquired with the cooperation of the Borobudur Conservation Office, the Nara National Research Institute for Cultural Properties, the Art Research Center of Ritsumeikan University, and the Asia-Japan Research Institute of Ritsumeikan University. We deeply thank these institutions for their generous cooperation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pavlidis, G.; Koutsoudis, A.; Arnaoutoglou, F.; Tsioukas, V.; Chamzas, C. Methods for 3D Digitization of Cultural Heritage. J. Cult. Herit. 2007, 8, 93–98. [Google Scholar]
  2. Pieraccini, M.; Guidi, G.; Atzeni, C. 3D Digitizing of Cultural Heritage. J. Cult. Herit. 2001, 2, 63–70. [Google Scholar]
  3. Li, R.; Luo, T.; Zha, H. 3D Digitization and Its Applications in Cultural Heritage. In Proceedings of the Euro-Mediterranean Conference, Lemesos, Cyprus, 8–13 November 2010; pp. 381–388. [Google Scholar]
  4. Barsanti, G.; Remondino, F.; Palacios, J.F. Critical Factors and Guidelines for 3D Surveying and Modelling in Cultural Heritage. Int. J. Herit. Digit. Era 2014, 3, 141–158. [Google Scholar]
  5. Kingsland, K. Comparative Analysis of Digital Photogrammetry Software for Cultural Heritage. Digit. Appl. Archaeol. Cult. Herit. 2020, 18, e00157. [Google Scholar] [CrossRef]
  6. Barsanti, S.G.; Guidi, G. A New Methodology for the Structural Analysis of 3D Digitized Cultural Heritage through FEA. IOP Conf. Ser. Mater. Sci. Eng. 2018, 364, 12005. [Google Scholar] [CrossRef]
  7. Doulamis, N.; Doulamis, A.; Ioannidis, C.; Klein, M.; Ioannides, M. Modelling of Static and Moving Objects: Digitizing Tangible and Intangible Cultural Heritage. In Mixed Reality and Gamification for Cultural Heritage; Springer: Berlin/Heidelberg, Germany, 2017; pp. 567–589. [Google Scholar]
  8. Wu, B.; Wan, A.; Yue, X.; Keutzer, K. Squeezeseg: Convolutional Neural Nets with Recurrent Crf for Real-Time Road-Object Segmentation from 3d Lidar Point Cloud. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; pp. 1887–1893. [Google Scholar]
  9. Zhou, D.; Fang, J.; Song, X.; Liu, L.; Yin, J.; Dai, Y.; Li, H.; Yang, R. Joint 3D Instance Segmentation and Object Detection for Autonomous Driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1839–1849. [Google Scholar]
  10. el Madawi, K.; Rashed, H.; el Sallab, A.; Nasr, O.; Kamel, H.; Yogamani, S. Rgb and Lidar Fusion Based 3d Semantic Segmentation for Autonomous Driving. In Proceedings of the 2019 IEEE Intelligent Transportation Systems Conference (ITSC), Auckland, NZ, USA, 27–30 October 2019; pp. 7–12. [Google Scholar]
  11. Biasutti, P.; Lepetit, V.; Aujol, J.-F.; Brédif, M.; Bugeau, A. Lu-Net: An Efficient Network for 3d Lidar Point Cloud Semantic Segmentation Based on End-to-End-Learned 3d Features and u-Net. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019; pp. 942–950. [Google Scholar]
  12. Bogoslavskyi, I.; Stachniss, C. Fast Range Image-Based Segmentation of Sparse 3D Laser Scans for Online Operation. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; pp. 163–169. [Google Scholar]
  13. Penza, V.; Ortiz, J.; Mattos, L.S.; Forgione, A.; de Momi, E. Dense Soft Tissue 3D Reconstruction Refined with Super-Pixel Segmentation for Robotic Abdominal Surgery. Int. J. Comput. Assist. Radiol. Surg. 2016, 11, 197–206. [Google Scholar] [PubMed]
  14. Su, Y.-H.; Huang, I.; Huang, K.; Hannaford, B. Comparison of 3d Surgical Tool Segmentation Procedures with Robot Kinematics Prior. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 4411–4418. [Google Scholar]
  15. Pierdicca, R.; Paolanti, M.; Matrone, F.; Martini, M.; Morbidoni, C.; Malinverni, E.S.; Frontoni, E.; Lingua, A.M. Point Cloud Semantic Segmentation Using a Deep Learning Framework for Cultural Heritage. Remote Sens. 2020, 12, 1005. [Google Scholar]
  16. Grilli, E.; Remondino, F. Classification of 3D Digital Heritage. Remote Sens. 2019, 11, 847. [Google Scholar]
  17. Matrone, F.; Grilli, E.; Martini, M.; Paolanti, M.; Pierdicca, R.; Remondino, F. Comparing Machine and Deep Learning Methods for Large 3D Heritage Semantic Segmentation. ISPRS Int. J. Geoinf. 2020, 9, 535. [Google Scholar]
  18. Mathias, M.; Martinovic, A.; Weissenberg, J.; Haegler, S.; van Gool, L. Automatic Architectural Style Recognition. ISPRS-Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2011, 3816, 171–176. [Google Scholar] [CrossRef]
  19. Oses, N.; Dornaika, F.; Moujahid, A. Image-Based Delineation and Classification of Built Heritage Masonry. Remote Sens. 2014, 6, 1863–1889. [Google Scholar] [CrossRef]
  20. Shalunts, G.; Haxhimusa, Y.; Sablatnig, R. Architectural Style Classification of Building Facade Windows. In Proceedings of the International Symposium on Visual Computing, Las Vegas, NV, USA, 26–28 September 2011; pp. 280–289. [Google Scholar]
  21. Zhang, L.; Song, M.; Liu, X.; Sun, L.; Chen, C.; Bu, J. Recognizing Architecture Styles by Hierarchical Sparse Coding of Blocklets. Inf. Sci. 2014, 254, 141–154. [Google Scholar] [CrossRef]
  22. Llamas, J.; Lerones, P.M.; Medina, R.; Zalama, E.; Gómez-García-Bermejo, J. Classification of Architectural Heritage Images Using Deep Learning Techniques. Appl. Sci. 2017, 7, 992. [Google Scholar]
  23. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep Learning on Point Sets for 3d Classification and Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  24. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. Adv. Neural. Inf. Process. Syst. 2017, 30, 5105–5114. [Google Scholar]
  25. Wu, W.; Qi, Z.; Fuxin, L. Pointconv: Deep Convolutional Networks on 3d Point Clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9621–9630. [Google Scholar]
  26. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic Graph Cnn for Learning on Point Clouds. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef]
  27. Thomas, H.; Qi, C.R.; Deschaud, J.-E.; Marcotegui, B.; Goulette, F.; Guibas, L.J. Kpconv: Flexible and Deformable Convolution for Point Clouds. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6411–6420. [Google Scholar]
  28. Hua, B.-S.; Tran, M.-K.; Yeung, S.-K. Pointwise Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 984–993. [Google Scholar]
  29. Zhao, H.; Jiang, L.; Jia, J.; Torr, P.H.S.; Koltun, V. Point Transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 16259–16268. [Google Scholar]
  30. Engel, N.; Belagiannis, V.; Dietmayer, K. Point Transformer. IEEE Access 2021, 9, 134826–134840. [Google Scholar] [CrossRef]
  31. Boulch, A.; le Saux, B.; Audebert, N. Unstructured Point Cloud Semantic Labeling Using Deep Segmentation Networks. 3DOR Eurograph. 2017, 3, 17–24. [Google Scholar]
  32. Lawin, F.J.; Danelljan, M.; Tosteberg, P.; Bhat, G.; Khan, F.S.; Felsberg, M. Deep Projective 3D Semantic Segmentation. In Proceedings of the International Conference on Computer Analysis of Images and Patterns, Ystad, Sweden, 22–24 August 2017; pp. 95–107. [Google Scholar]
  33. Tatarchenko, M.; Park, J.; Koltun, V.; Zhou, Q.-Y. Tangent Convolutions for Dense Prediction in 3d. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3887–3896. [Google Scholar]
  34. Zhou, H.; Zhu, X.; Song, X.; Ma, Y.; Wang, Z.; Li, H.; Lin, D. Cylinder3d: An Effective 3d Framework for Driving-Scene Lidar Semantic Segmentation. arXiv 2020, arXiv:2008.01550. [Google Scholar]
  35. Cheng, R.; Razani, R.; Taghavi, E.; Li, E.; Liu, B. Af2-S3net: Attentive Feature Fusion with Adaptive Feature Selection for Sparse Semantic Segmentation Network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12547–12556. [Google Scholar]
  36. Xu, J.; Zhang, R.; Dou, J.; Zhu, Y.; Sun, J.; Pu, S. Rpvnet: A Deep and Efficient Range-Point-Voxel Fusion Network for Lidar Point Cloud Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 16024–16033. [Google Scholar]
  37. Yan, X.; Gao, J.; Zheng, C.; Zheng, C.; Zhang, R.; Cui, S.; Li, Z. 2dpass: 2d Priors Assisted Semantic Segmentation on Lidar Point Clouds. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2022; pp. 677–695. [Google Scholar]
  38. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  39. Iandola, F.; Moskewicz, M.; Karayev, S.; Girshick, R.; Darrell, T.; Keutzer, K. Densenet: Implementing Efficient Convnet Descriptor Pyramids. arXiv 2014, arXiv:1404.1869. [Google Scholar]
  40. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. Imagenet: A Large-Scale Hierarchical Image Database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  41. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [PubMed]
  42. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
  43. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected Crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar]
  44. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  45. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  46. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  47. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  48. Rusu, R.B. Semantic 3D Object Maps for Everyday Robot Manipulation; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  49. Demantké, J.; Mallet, C.; David, N.; Vallet, B. Dimensionality Based Scale Selection in 3D Lidar Point Clouds. In Laserscanning; HAL Open Science: Calgary, AB, Canada, 2011. [Google Scholar]
  50. Mallet, C.; Bretar, F.; Roux, M.; Soergel, U.; Heipke, C. Relevance Assessment of Full-Waveform Lidar Data for Urban Area Classification. ISPRS J. Photogramm. Remote Sens. 2011, 66, S71–S84. [Google Scholar]
  51. Weinmann, M.; Jutzi, B.; Mallet, C. Semantic 3D Scene Interpretation: A Framework Combining Optimal Neighborhood Size Selection with Relevant Features. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2014, 2, 181. [Google Scholar] [CrossRef] [Green Version]
  52. He, E.; Chen, Q.; Wang, H.; Liu, X. A curvature based adaptive neighborhood for individual point cloud classification. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 42, 219–225. [Google Scholar] [CrossRef]
  53. Kawakami, K.; Hasegawa, K.; Li, L.; Nagata, H.; Adachi, M.; Yamaguchi, H.; Thufail, F.I.; Riyanto, S.; Tanaka, S. Opacity-based edge highlighting for transparent visualization of 3d scanned point clouds. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2020, 5, 373–380. [Google Scholar] [CrossRef]
  54. Tanaka, S.; Hasegawa, K.; Okamoto, N.; Umegaki, R.; Wang, S.; Uemura, M.; Okamoto, A.; Koyamada, K. See-through imaging of laser-scanned 3d cultural heritage objects based on stochastic rendering of large-scale point clouds. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 3, 73–80. [Google Scholar] [CrossRef] [Green Version]
  55. Uchida, T.; Hasegawa, K.; Li, L.; Adachi, M.; Yamaguchi, H.; Thufail, F.I.; Riyanto, S.; Okamoto, A.; Tanaka, S. Noise-Robust Transparent Visualization of Large-Scale Point Clouds Acquired by Laser Scanning. ISPRS J. Photogramm. Remote Sens. 2020, 161, 124–134. [Google Scholar]
  56. Girardeau-Montaut, D. CloudCompare. 2016. Available online: http://www.cloudcompare.org/ (accessed on 6 February 2023).
  57. Cignoni, P.; Callieri, M.; Corsini, M.; Dellepiane, M.; Ganovelli, F.; Ranzuglia, G. Meshlab: An Open-Source Mesh Processing Tool. In Proceedings of the Eurographics Italian Chapter Conference, Salerno, Italy, 2–4 July 2008; Volume 2008, pp. 129–136. [Google Scholar]
  58. Jutzi, B.; Gross, H. Nearest Neighbour Classification on Laser Point Clouds to Gain Object Structures from Buildings. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2009, 38, 4–7. [Google Scholar]
  59. Li, X.; Li, X.; Zhang, L.; Cheng, G.; Shi, J.; Lin, Z.; Tan, S.; Tong, Y. Improving Semantic Segmentation via Decoupled Body and Edge Supervision. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 435–452. [Google Scholar]
  60. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
Figure 1. Borobudur Temple in Indonesia (photograph). (a) Borobudur temple; (b) corridors with 1460 narrative relief panels on walls. The total relief surface is 2500 square meters.
Figure 1. Borobudur Temple in Indonesia (photograph). (a) Borobudur temple; (b) corridors with 1460 narrative relief panels on walls. The total relief surface is 2500 square meters.
Remotesensing 15 00956 g001
Figure 2. Borobudur Temple Relief, Karmawibhangga, panel No. 22, tells the story of a group of men carrying umbrellas who approached a holy man sitting in the shade of a wooden building, while people nearby were peacefully chatting. To tell these beautiful Buddhist stories, the reliefs are carved with different patterns that are raised only slightly from the background. The reliefs can be divided into several semantic labels: characters, plants, animals, ships, buildings, and so on.
Figure 2. Borobudur Temple Relief, Karmawibhangga, panel No. 22, tells the story of a group of men carrying umbrellas who approached a holy man sitting in the shade of a wooden building, while people nearby were peacefully chatting. To tell these beautiful Buddhist stories, the reliefs are carved with different patterns that are raised only slightly from the background. The reliefs can be divided into several semantic labels: characters, plants, animals, ships, buildings, and so on.
Remotesensing 15 00956 g002
Figure 3. Block diagram: The first step is to extract the relief panel from the input and project the RGB image and depth map. Then, the soft edge is extracted by opacity-based edge highlighting. These images are transmitted to the soft-edge enhanced network as a multichannel to predict the two-dimensional semantic segmentation labels. Finally, the results are obtained by mapping the extracted semantic labels onto three-dimensional points of the relief data.
Figure 3. Block diagram: The first step is to extract the relief panel from the input and project the RGB image and depth map. Then, the soft edge is extracted by opacity-based edge highlighting. These images are transmitted to the soft-edge enhanced network as a multichannel to predict the two-dimensional semantic segmentation labels. Finally, the results are obtained by mapping the extracted semantic labels onto three-dimensional points of the relief data.
Remotesensing 15 00956 g003
Figure 4. Relief panel extraction.
Figure 4. Relief panel extraction.
Remotesensing 15 00956 g004
Figure 5. Soft-edge extraction from relief photogrammetric point cloud by opacity-based edge highlighting and image projection.
Figure 5. Soft-edge extraction from relief photogrammetric point cloud by opacity-based edge highlighting and image projection.
Remotesensing 15 00956 g005
Figure 6. Typical relation between the three-dimensional feature value f and opacity α when adopting the function.
Figure 6. Typical relation between the three-dimensional feature value f and opacity α when adopting the function.
Remotesensing 15 00956 g006
Figure 7. Structure of the soft-edge enhanced network. This is a multichannel input network with RGB images, depth maps, and soft-edge images. The network utilizes DenseNet121 as the backbone and extracts five sizes of feature layers: H/2, H/4, H/8, H/16, and H/32. Three decoder blocks (as shown in the upper right) are set to supervise the segmentation edge and segmentation body intermediate layer and semantic segmentation results in the decoder sections.
Figure 7. Structure of the soft-edge enhanced network. This is a multichannel input network with RGB images, depth maps, and soft-edge images. The network utilizes DenseNet121 as the backbone and extracts five sizes of feature layers: H/2, H/4, H/8, H/16, and H/32. Three decoder blocks (as shown in the upper right) are set to supervise the segmentation edge and segmentation body intermediate layer and semantic segmentation results in the decoder sections.
Remotesensing 15 00956 g007
Figure 8. Block diagram of dataset creation.
Figure 8. Block diagram of dataset creation.
Remotesensing 15 00956 g008
Figure 9. Predicted results on test data. Left to right: input RGB image, depth map, soft-edge, ground truth, and prediction results. The results show that our network is robust to different concentrations in soft edges.
Figure 9. Predicted results on test data. Left to right: input RGB image, depth map, soft-edge, ground truth, and prediction results. The results show that our network is robust to different concentrations in soft edges.
Remotesensing 15 00956 g009
Figure 10. Prediction for segmentation edges. Left to right: ground truth of the segmentation edge, prediction results of the segmentation edge, ground truth of semantic segmentation, and prediction results of semantic segmentation. The results show that the network can predict approximate segmentation edges and is consistent with the prediction results of semantic segmentation.
Figure 10. Prediction for segmentation edges. Left to right: ground truth of the segmentation edge, prediction results of the segmentation edge, ground truth of semantic segmentation, and prediction results of semantic segmentation. The results show that the network can predict approximate segmentation edges and is consistent with the prediction results of semantic segmentation.
Remotesensing 15 00956 g010
Figure 11. Left to right: input RGB image, ground truth of semantic segmentation, prediction results of our proposed network, and prediction results of SegNet, U-Net, PSPNet, and DeepLabV3+. The results show that extant networks are more sensitive to the color information of input images and demonstrate the robustness of our proposed network to photo information.
Figure 11. Left to right: input RGB image, ground truth of semantic segmentation, prediction results of our proposed network, and prediction results of SegNet, U-Net, PSPNet, and DeepLabV3+. The results show that extant networks are more sensitive to the color information of input images and demonstrate the robustness of our proposed network to photo information.
Remotesensing 15 00956 g011
Figure 12. Visualization results of mapping two-dimensional prediction results to three-dimensional point cloud.
Figure 12. Visualization results of mapping two-dimensional prediction results to three-dimensional point cloud.
Remotesensing 15 00956 g012
Figure 13. Samples of visualization results.
Figure 13. Samples of visualization results.
Remotesensing 15 00956 g013
Figure 14. Visualization results of applying our proposed semantic segmentation framework on photogrammetric point cloud data.
Figure 14. Visualization results of applying our proposed semantic segmentation framework on photogrammetric point cloud data.
Remotesensing 15 00956 g014
Table 1. Confusion matrix and evaluation metrics of the network on test data.
Table 1. Confusion matrix and evaluation metrics of the network on test data.
MatrixBackgroundCharactersPlantsOthersRecallPrecisionIoUF1-Score
Background6,036,161100,86770,766443,8410.90750.89830.82300.9029
Characters129,2607,635,22619,573114,3880.96670.93760.90830.9519
Plants144,16863,7122,689,93834,0970.91750.89310.82670.9051
Others409,643343,360231,6053,771,0910.79300.86430.70510.8271
MeanAccuracy: 0.90530.89610.89830.81580.8968
Table 2. Comparison of evaluation metric results between current networks and our proposed network. The results demonstrate that our network outperforms other networks in all metrics.
Table 2. Comparison of evaluation metric results between current networks and our proposed network. The results demonstrate that our network outperforms other networks in all metrics.
NetworkRecallPrecisionmIoUF1-ScoreAccuracy
SegNet0.68430.70790.53880.69320.7240
U-Net0.67080.70840.52510.68390.7120
PSPNet0.69530.70960.54700.70110.7184
DeeplabV3+0.67070.68770.52000.67770.6993
Ours0.89610.89830.81580.89680.9053
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ji, S.; Pan, J.; Li, L.; Hasegawa, K.; Yamaguchi, H.; Thufail, F.I.; Brahmantara; Sarjiati, U.; Tanaka, S. Semantic Segmentation for Digital Archives of Borobudur Reliefs Based on Soft-Edge Enhanced Deep Learning. Remote Sens. 2023, 15, 956. https://doi.org/10.3390/rs15040956

AMA Style

Ji S, Pan J, Li L, Hasegawa K, Yamaguchi H, Thufail FI, Brahmantara, Sarjiati U, Tanaka S. Semantic Segmentation for Digital Archives of Borobudur Reliefs Based on Soft-Edge Enhanced Deep Learning. Remote Sensing. 2023; 15(4):956. https://doi.org/10.3390/rs15040956

Chicago/Turabian Style

Ji, Shenyu, Jiao Pan, Liang Li, Kyoko Hasegawa, Hiroshi Yamaguchi, Fadjar I. Thufail, Brahmantara, Upik Sarjiati, and Satoshi Tanaka. 2023. "Semantic Segmentation for Digital Archives of Borobudur Reliefs Based on Soft-Edge Enhanced Deep Learning" Remote Sensing 15, no. 4: 956. https://doi.org/10.3390/rs15040956

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop