Next Article in Journal
A2PM-VINS: A Visual–Inertial SLAM Method Based on Area-to-Point Matching
Previous Article in Journal
Label-Free Optical Sensor for Real-Time Monitoring of Insulin Secretion from Single Human Pancreatic Islets
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

FOR: Point Cloud Outlier Removal Based on Fuzzy Theory and Informativeness and Its Application to 3D Object Detection

School of Big Data and Software Engineering, Chongqing University, Chongqing 401331, China
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(10), 3070; https://doi.org/10.3390/s26103070
Submission received: 29 March 2026 / Revised: 22 April 2026 / Accepted: 30 April 2026 / Published: 13 May 2026
(This article belongs to the Special Issue Recent Progress in 3D Computer Vision and Robotics)

Abstract

LiDAR is widely used in autonomous driving. Although LiDAR point cloud data can provide stable and reliable information about the environment, it also faces the problem of a huge amount of data. One of the reasons is that point cloud data contains a large amount of noise and outliers. Outlier removal of point clouds can reduce the impact of these disturbances and improve the quality of the point cloud, but it will inevitably eliminate some valid points, which affects subsequent perception tasks. To overcome this limitation, this paper proposes a fuzzy outlier removal (FOR) method based on fuzzy theory and informativeness. It uses fuzzy theory to model the uncertainty of the membership degree of each point in each dimension, calculates the informativeness sum of each point based on membership degree, and filters points according to the informativeness. FOR is characterized by filtering the point cloud in the edge region on the premise of retaining the point cloud in the center region, so as to preserve the environmental information in the center region and reduce the impact of outlier removal on subsequent perception tasks. The experiments focus on the contradictory relationship between outlier removal and perception accuracy, and verify the effectiveness of FOR with multiple object detection models on the autonomous driving datasets KITTI and nuScenes. The experimental results indicate that, compared with other point cloud outlier removal methods, FOR has the advantage of reducing inference time while retaining detection accuracy, demonstrating balanced high performance across different datasets and detection models.

1. Introduction

With the continuous advancement of LiDAR technology and cost reduction, LiDAR has been used as an advanced environment-sensing sensor and applied to many fields such as autonomous driving, aerospace, terrain survey, etc. [1,2,3]. LiDAR mainly acquires point clouds by transmitting and receiving electromagnetic waves to realize the survey of the environment. High-resolution LiDAR can obtain hundreds of thousands of point clouds in each frame, providing an adequate representation of the scene.
Due to the limitations of sensor and measurement errors, point cloud data often contain a large amount of noise and large number of outliers, which not only brings great challenges for data storage and representation and computational resource consumption, but also hinders the application of downstream geometric processing [4,5]. Therefore, point cloud outlier removal has become one of the key technologies of preprocessing. The reasons for outlier generation include the physical properties of LiDAR, in addition to factors such as illumination and natural environment. As the laser of the LiDAR is emitted from the center to the surrounding 360°, the received point cloud shows an uneven density distribution. The points in the regions farther away from the center are sparser, and usually lose the geometric features of the object and become outlier points. Vehicle-mounted LiDAR can sense obstacles up to 500 m away, while the edge range contains more outliers (as shown in Figure 1). At the same time, targets closer to the center on actual traffic roads are of more concern, and existing detection techniques are more efficient for closer targets. This indicates that an outlier removal technique specialized for long-range noise is needed to clean the point cloud and improve its quality.
Three-dimensional object detection is an important branch in autonomous driving, and many LiDAR-based 3D object detection techniques have emerged in recent years and made significant breakthroughs in this direction. Unfortunately, these studies have rarely considered outlier removal in point clouds, and a few studies have only focused on point cloud outlier removal under extreme weather [6,7]. Limited by the physical properties of LiDAR, as the distance increases, the intensity of the laser signal will be attenuated due to atmospheric absorption and scattering, resulting in a weaker echo signal and a sparser point cloud in the distance. In addition, the echo signals in the distance may be subject to more noise interference, which generates more outliers in the point cloud.
There are two main reasons why existing point cloud outlier removal methods are seldom applied in 3D object detection:
(1) Limitations of local feature analysis. Existing methods such as ROR and SOR mainly rely on local features to discriminate outliers, lacking consideration of the global spatial position of points. ROR judges a point as an outlier if the number of neighbors within a fixed radius falls below a threshold; SOR treats points deviating from the local mean distance by more than a standard deviation ratio as outliers. Because these methods only analyze local geometric properties without incorporating the overall spatial distribution, they struggle to effectively remove distant and dense noise points.
(2) Binary classification defects. Traditional methods adopt binary (hard) thresholding for noise discrimination, which ignores the continuous and uncertain nature of noise intensity. This rigid decision is prone to misclassifying valid points in the center region, resulting in the loss of critical valid points and seriously degrading 3D object detection performance. This creates a fundamental contradiction between outlier removal and detection accuracy.
To address the above problems, the main contributions of this paper are as follows:
(1) A fuzzy informativeness-driven outlier removal algorithm (FOR). We propose a novel filtering method based on fuzzy theory and informativeness. The spatial position of each point is mapped to a three-dimensional fuzzy membership vector, reflecting the confidence probability of belonging to the real target surface. A weighted logarithmic informativeness model is constructed to quantify the uncertainty degree (i.e., noise likelihood) of each point. Based on global informativeness ranking, the top k% high-informativeness points are dynamically truncated, filtering distant edge-region noise while maximally preserving center-region object features.
(2) Validation on public datasets. Based on multiple advanced detection models on KITTI and nuScenes, we explore the contradictory relationship between outlier removal and detection performance. The results show that FOR achieves optimal performance on KITTI. On nuScenes, though slightly lower than NiOR and SOR, FOR still outperforms DROR and ROR, exhibiting stable cross-platform adaptability.
The rest of the paper is organized as follows: the second part reviews the related research in this paper, and the third part proposes FOR based on fuzzy theory and informativeness. Then, in the fourth part, relevant experiments are conducted to verify the effectiveness of FOR. The last part summarizes the content of this paper.

2. Related Work

In recent years, many studies have centered around point cloud outlier removal, from filtering-based methods to deep learning-based methods [8,9]. Among the traditional filter-based outlier removal methods, the classical methods are radius outlier removal (ROR) and statistical outlier removal (SOR). Among them, ROR considers the density within the neighborhood of a point, and if the density of the neighborhood is too small, the point is regarded as an outlier. SOR performs a statistical analysis of the neighborhood of each point and filters out some outliers that do not meet the requirements based on the features of the distribution of distances from the point to all neighboring points. Many studies have optimized these two methods and proposed new methods in research on ROR. Duan et al. [10] proposed an adaptive radius outlier removal filter based on principal component analysis (PCA), which used PCA to downsize the point cloud to get 2D point cloud data in order to reduce the time complexity of the algorithm, and then used the 2D principal components to adaptively adjust the neighborhood search radius of each point. Similarly, Makhluk et al. [11] used the Euclidean distance from the point to the coordinate origin and additional parameters to dynamize the search neighborhood of each point, and the algorithm had a good filtering effect in extreme weather. Szutor et al. [12], for traditional ROR, needed to carry out the nearest-neighbor search many times, which led to an increase in computation, combined with the idea of the cell-based filter to improve the time performance of the algorithm. In studies on SOR, Haris et al. [13] extended traditional SOR by fast clustering statistical filtering, which clustered the input point cloud, filtered the clusters with higher density by taking the average value of all clusters’ points as the threshold, and then filtered the remaining point cloud by the traditional SOR method. This method sped up the computation by being the first to retain the denser part of the point cloud through clustering and reduced the number of points and clusters actually estimated. Carrilho et al. [14] proposed an adaptive scheme based on the cellular subdivision of the point cloud. Instead of computing a single histogram for the entire dataset, the method filtered smaller patches so that ground elevation differences were ignored.
Deep learning-based point cloud outlier removal methods focus on learning and representing the features of outliers in the learning stage and identifying the points with similar features in the running stage. Rakotosaona et al. [15] proposed a method for removing outliers and reducing noise based on the structure of PCPNet [16], which utilized a network model to directly classify and discard outlier points. Pistilli et al. [17] proposed an outlier removal method based on graph convolutional neural networks, which dynamically constructed a neighborhood graph based on the similarity of high-dimensional feature representations of points, constructed a complex hierarchy of features, and utilized a convolutional layer to classify and discard the outliers. Li et al. [18] proposed a single-stage point cloud cleaning network (SSPCN) for implementing outlier filtration and denoising of point clouds, in which an improved density-based farthest point sampling (DBFPS) method was used for downsampling and outlier removal.
Deep learning-based outlier removal methods need to represent the high-dimensional features of the point cloud, ignoring the basic location information of point clouds, so most studies still focus on filtering-based outlier removal algorithms [19,20,21,22]. In addition, there are some novel outlier removal methods. For example, Li et al. [23] proposed PointCVaR, which defined the concept of point risk and reshaped the outlier removal process into an optimization problem using gradient-based attribution analysis.
Existing research methods do not focus on the removal of outliers from the point cloud, and fail to find a balance between detection time and detection accuracy, always paying too much attention to one side and ignoring the other. The FOR proposed in this paper will maximize the retention of point cloud features in the central region of the scene under the premise of filtering the distant noise and outliers and ensure detection accuracy while reducing the detection time and improving the detection performance.

3. Method

Aiming to address the limitations of traditional filtering-based point cloud outlier removal methods in 3D object detection, we propose FOR based on fuzzy theory and informativeness to filter the point cloud from a distance. Since distance is a precise quantitative metric, but ’near’ and ’far’ are fuzzy concepts in reality, fuzzy theory is used to model the uncertainty of the membership of points in the central region in order to better portray these fuzzy concepts.

3.1. Fuzzy Theory and Informativeness

Fuzzy theory. Fuzzy set theory is a tool used to represent uncertainty and ambiguity, i.e., to express the assessment of things in terms of numbers between 0 and 1. In recent years, the concept of fuzzy set theory has been gradually introduced into point cloud data processing and 3D object detection [24,25,26]. Fuzzy set theory is defined as follows:
For a given mapping over the discussion domain, as in Equation (1),
μ A : X [ 0 , 1 ]
where A is a fuzzy subset on the discussion domain X, μ A ( ) is the membership function of A, and μ A ( x ) is the degree to which x is attributed to A, expressed as in Equation (2):
A = { ( x , μ A ( x ) ) | x X }
As in a point cloud, the magnitude of the degree of membership of a point P on the x-axis to the entire point cloud data is indicated μ X ( P ) . A common way to find the degree of membership is to simplify the membership function into a triangle. Let A be a subset of fuzzy numbers on the set of real numbers R. Its membership function μ A is a continuous mapping from the real number domain to the interval [0, 1].
The membership function of a triangular fuzzy set is calculated as follows:
μ A ( x ) = { ( x c + δ ) / ( a c + δ ) , c x a ( x b δ ) / ( a b δ ) , a x b .
where c and b denote the lower and upper limits of the assessment variables, respectively, a denotes the intermediate value, and the average of c and b is usually taken. As x gets closer to a, μ A ( x ) gets closer to 1, indicating higher membership of x to A. Conversely, the closer μ A ( x ) is to 0, the lower the membership of x to A. δ is usually used to adjust the range of μ A ( x ) so that it is in (0, 1]. In the point cloud data, the minimum value of c, the maximum value of b, and the average value of a are found for all the point clouds on the x-axis. For the value of the x-axis of each point cloud, the closer the points are to c and b, the more likely they are to be edge points and noisy points, and the closer the points are to a, the more likely they are to be points near the center. The same is true for the y-axis and the z-axis. In this paper, δ is set to δ = (b − c)/n, where n denotes the number of evaluation variables, i.e., the number of points in the space.
For the probability of a single event, the uncertainty can be measured in terms of the informativeness, as in Equation (4). Informativeness H is a concept based on informativeness theory and can be intuitively interpreted as the ‘degree of unexpectedness’ or ‘uncertainty’. Compared with the degree of membership, the scale is unified, which can be better analyzed.
H = l o g p
It is easy to prove that the function curve of Equation (4) is monotonically decreasing in the interval (0, 1], i.e., the higher the probability for a single event, the less informative it is and the less uncertain it is. Conversely, the smaller the probability, the more informative and the greater the uncertainty. Therefore, informativeness can be used to measure the uncertainty of the proposition of whether the point cloud belongs to the center region. Frankly speaking, informativeness H(p) = −log(p) measures the ’unexpected degree’ of the point cloud location: the informativeness of the central region point (high p) is low, because its location is expected to be clear; edge points (low p) contain a lot of informativeness, which may be outliers or invalid data. Unlike the probability estimation, the location of the point cloud is certain in space, and the probability is expressed by the membership degree of the point cloud with the center region, which is also certain, and the uncertainty of this proposition should be as small as possible.
In consideration of the additivity of the informativeness, the sum of the informativeness of each point can be computed and weighted according to its membership degree on the x, y, and z axes, as in Equation (5).
E i = α l o g μ X ( P i ) β l o g μ Y ( P i ) γ l o g μ Z ( P i )
where μ j ( P i ) , j∈{X,Y,Z} denotes the membership of the ith point to each coordinate axis and α, β, γ denote the weight of each dimension, respectively. FOR aims to filter the outlier points at a distance, so focusing on the distribution of the point cloud in the XOY plane, the computation and experiments in this paper set the weights as α = 0.4, β = 0.4, γ = 0.2.
FOR combines fuzzy theory and informativeness metrics, skillfully using membership degree to model the uncertainty of point clouds and screen out the most likely point clouds in the central region. The overall framework of FOR is shown in Figure 2.

3.2. FOR Algorithm

FOR calculates the membership of each point cloud on different axes with respect to the center region. It considers the membership as a probability and calculates the sum of the amount of informativeness of each point cloud on each axis. The membership obtained by the points in the edge region is low, and the sum of the informativeness is large. After calculating the sum of the informativeness of each point, the points are sorted by the size of informativeness. According to the set outlier ratio k (e.g., 0.2), which indicates that 20% of the points will be removed, the points with the top 20% of the total informativeness will be filtered out. The specific steps are shown in Algorithm 1.
Algorithm 1 FOR
Input: Points                         - A set of point clouds to be filtered
k                         - Outlier ratio
Output: denosiedPoints                     - Point cloud after outlier removal

# Calculate fuzzy number
1     fuzzy_num = [min(points, axis = 0),0, max(points, axis = 0)]
2     entropy_list = []
3     for p points do
# Calculate the membership of each point
4       membership = Membership(fuzzy_num, p)
# Calculating informativeness
5       e = CalculateInformativeness(membership)
6       informativeness_list.append(e)
7     end for
8     Sort(informativeness_list)
# Find the informativeness threshold
9     threshold = Search(informativeness_list, k)
10  for all i do
11     if informativeness_list [i] > threshold:
12         denosiedPoints = Filter(points, points[i])
13     end if
14  end for
15  return denosiedPoints
In order to illustrate the flow of the algorithm more intuitively, the details of the algorithm are demonstrated through an arithmetic example.
Step 1: Let the space include 10,000 points, and construct the triangular fuzzy number of the x, y, and z axes based on the coordinate values of all point clouds. As:
x_- min = −77.28; x_- mid = 0; x_- max = 78.69;
y_- min = −67.55; y_- mid = 0; y_- max = 55.48;
z_- min = −9.56; z_- mid = 0; z_- max = 2.78.
Step 2: Calculate the degree of membership of each point on each axis using Equation (2), e.g., the coordinates of a point P are (−17.5, 18, −0.9). Then, its degree of membership on each of the three coordinate axes is μ_X (P) = 0.773, μ_Y (P) = 0.672, μ_Z (P) = 0.905.
Step 3: Weigh the sum of the informativeness of the point according to Equation (5) to obtain E_P = 0.122.
Step 4: Find the informativeness of each point and sort it by size. According to the set proportion of the outlier value, remove the points outside the range.
The actual filtering effect of FOR is shown in Figure 3 and compared with the other two commonly used outlier removal algorithms through the parameters of each algorithm, so that the difference between the number of points they retained is within 1000. In this case, the effect of FOR is similar to that of the other methods, which verifies the validity of FOR and the two outlier removal methods, SOR and ROR. There are still some noise points in the range of the edges that cannot be filtered. At the same time, FOR focuses more on retaining the point cloud in the center region, so the distribution of the filtered point cloud is more concentrated. Compared with other methods, it can be adjusted by fewer parameters. The effect of FOR on the object detection results and its performance analysis will be discussed in detail in Section 4, Experiment.

3.3. Time Complexity

The time complexity statistics of FOR are shown in Table 1. The time complexity of FOR mainly focuses on the step of sorting the informativeness for each point, so its time complexity is N*(logN). SOR and ROR, on the other hand, mainly traverse each point and search for the neighboring points of each point, in which the search process is realized by using a spatial partitioning search structure on the basis of kd-tree, with a complexity of O(logN) and an overall time complexity of O(N*logN). Thus, FOR has the same time complexity as SOR and ROR.

4. Experiment

4.1. Experimental Setup

In order to evaluate the effectiveness of FOR, validation is performed on the autonomous driving open-source dataset using several 3D object detection models. The experiment is conducted on the validation sets of the KITTI [27] and nuScenes [28] datasets, with the main objective of exploring the accuracy loss of the data after FOR under different detection models. In terms of data sample distribution, the validation set of the KITTI dataset (64-line LiDAR) contains 3769 frames of point cloud data, which mainly detects the three categories of cars, pedestrians, and cyclists. The validation set of the nuScenes dataset (32-line LiDAR) contains 150 scenes of continuous point cloud data, which mainly detects 10 categories such as cars, pedestrians, and trucks. In terms of evaluation metrics, the KITTI dataset is evaluated in four aspects: 3D bounding box, 2D bounding box, bird’s-eye view (BEV), and average orientation similarity (AOS).
Three-dimensional: The accuracy of the 3D bounding box of an object is predicted using the IOU as a reference standard, and if the IOU is greater than a set threshold, the prediction is a positive sample.
Two-dimensional: The predicted frame is projected to the camera view through the camera parameters, and its accuracy is calculated in the 2D plane, again judging positive and negative samples based on the IOU.
BEV: The prediction frame is projected to the bird’s-eye view perspective to calculate its accuracy, and the positive and negative samples are judged according to the IOU.
AOS: This is an accuracy index that indicates the orientation of the prediction frame. Unlike 2D detection, 3D object detection usually needs to pay attention to the orientation information of the target. The AOS is expressed by the cosine similarity (OS) of the two direction vectors of the angle of the prediction frame (θ_pred) and the angle of the real frame (θ_gt), as in Equation (6), and then the OS of each category is averaged to obtain the average cosine similarity (AOS).
O S = | c o s ( θ p r e d θ g t ) |
In addition, two metrics, AP11 and AP40, can be considered depending on the different settings of the accuracy thresholds. The nuScenes dataset defines an error metric mTP = {mATE, mASE, mAOE, mAVE, mAAE}, in addition to the average accuracy of detecting the bounding box, and a comprehensive evaluation metric, NDS, the details of which can be viewed in [28].
In order to get more comprehensive evaluation results, the selected detection models should all differ in terms of algorithms, which are MVF [29], MVXNet [30], Part-A2 [31], and PointPillars [32]. Among them, MVF employs a multi-view fusion detection method, MVXNet is a classic in multi-modal fusion detection schemes, and Part-A2 introduces partial perception and aggregation. PointPillars, on the other hand, employs a pillar-based encoding of point cloud features. All detectors (MVF, MVXNet, Part-A2, PointPillars) use publicly available pre-trained weights provided by the original authors or the MMDetection3D framework. Table 2 shows the same configuration used in the experiment.

4.2. Experimental Results

In addition to the classical ROR and SOR, the Dynamic Radius Outlier Removal Filter (DROR) [7] proposed by Charron et al. and the Non-isolated outlier removal (NiOR) proposed by Ning et al. [33] are also referenced. Specifically, the parameters of each filtering method are adjusted so that the number of retained points differs by less than 1000 points per frame compared to FOR-0.25 (which retains approximately 75% of original points). This principle is applied consistently in both the qualitative visualization (Figure 3) and all quantitative comparisons on KITTI and nuScenes.
The validation results of the KITTI dataset on different models after different outlier removal methods are shown in Figure 4, Figure 5, Figure 6 and Figure 7, which record the precision of the three models under four metrics: 2D bounding box, 3D bounding box, BEV bounding box, and AOS, where the precision values are the average of the precision of the three categories (cars, pedestrians, and cyclists) and are partitioned according to the recall points (AP11 and AP40). The specific mAP values in Figure 4, Figure 5, Figure 6 and Figure 7 are shown in Table A1, Table A2, Table A3 and Table A4 in Appendix A. The outlier ratio for FOR (orange in Figure 4, Figure 5, Figure 6 and Figure 7) is set to 0.25, indicating that 25% of the points in each frame of data are filtered out. The ROR (gray in Figure 4, Figure 5, Figure 6 and Figure 7) parameter uses the default parameters radius = 0.5 and number of neighborhood points = 10. The default parameters for SOR (purple in Figure 4, Figure 5, Figure 6 and Figure 7) are standard deviation ratio = 0.5 and number of neighborhood points = 10. For DROR (blue in Figure 4, Figure 5, Figure 6 and Figure 7), the number of neighborhood points is 10 and the minimum radius is 0.5. For NiOR (pink in Figure 4, Figure 5, Figure 6 and Figure 7), the number of neighborhood is 10 and the threshold is 0.55. By analyzing Figure 4, Figure 5, Figure 6 and Figure 7, it can be concluded that the accuracy of each outlier removal method on different models is, in descending order, FOR > DROR > ROR > SOR > NiOR.
After filtering some points from the original data, which include some points of objects in addition to noise, the accuracy of the filtered data is all reduced compared to the original data. However, in each metric, the accuracy of FOR is closest to the accuracy of the original data, and the error in the accuracy of each metric compared to the original data is around 5%, which is much smaller than the error of the other four outlier removal methods. This is because FOR gives priority to ensuring the point cloud within the center neighborhood from the overall point cloud density, and the center region is the region with the highest point cloud density and the most obvious object features. Therefore, FOR effectively protects the features within this region, thus minimizing the object detection error. Other outlier removal methods filter from local density without considering the problem of excessive density in the center region, which leads to the loss of more effective points, and, thus, the accuracy is seriously degraded.
Table 3 shows the validation results of Pointpillars on the nuScenes-trainval dataset after processing with different outlier removal methods, and according to the comprehensive evaluation metric NDS, the comprehensive performance of each outlier removal method, in descending order, is as follows: NiOR> SOR> FOR> ROR> DROR. Combined with their performance on KITTI, NiOR and SOR perform well on nuScenes but decline significantly on KITTI; DROR and ROR show the opposite trend. In contrast, FOR achieves optimal performance on KITTI. On nuScenes, though slightly lower than NiOR and SOR, FOR still outperforms DROR and ROR, exhibiting no significant performance fluctuations across different sensor configurations (64-line/32-line LiDAR), demonstrating good cross-platform adaptability and robustness.
Table 4 shows the FPS values of different data under different models, and it can be seen that the FPS after different outlier removal is higher than the original FPS because the filtering reduces the number of original point clouds. FOR achieves the highest FPS on KITTI across all models, with a 124% improvement on Part-A2. On nuScenes, FOR is sub-optimal, second only to DROR.
Combining accuracy (Table 3) and speed (Table 4) results reveals distinct method characteristics. DROR achieves the fastest inference but suffers from the lowest NDS, indicating an aggressive outlier removal strategy that sacrifices accuracy for speed. Conversely, NiOR achieves the highest NDS but shows limited FPS improvement, suggesting conservative outlier removal that preserves accuracy at the cost of efficiency. FOR occupies an optimal balance: on KITTI, it dominates both metrics (highest accuracy and highest FPS); on nuScenes, its FPS is competitive (second only to DROR), while its NDS remains close to the best performers (NiOR). This demonstrates FOR’s superior ability to simultaneously optimize accuracy and efficiency, avoiding the single-metric extremes of other methods.
Table 5 shows the detection results of FOR with different outlier ratio settings. From the table, we can analyze that on each index of each model, the detection accuracy decreases with an increase in the outlier ratio. This is due to the fact that with an increase in the outlier ratio, more points are filtered, which results in affecting the accuracy of the detection. However, the accuracies at different outlier ratios are very close to each other, and the error between each noise point ratio and the neighboring noise point ratios is within 1%, which indicates that the parameter settings of FOR have a small impact on the detection. According to the statistics, the average effective area of each frame in the KITTI dataset consists of 18,560 points, and the difference between different noise point ratios is 5%, which means that each time we adjust the parameters, we will filter out about 928 more points, and these 928 points are filtered from the farthest range, which will not affect the objects in the center area and minimize the loss of accuracy.

5. Conclusions

As a fundamental environment perception sensor, LiDAR generates precise 3D representations of real-world scenes through point cloud data, with widespread applications in autonomous driving and related fields. However, conventional outlier removal processes inevitably eliminate valid points during noise suppression, adversely affecting subsequent 3D object detection performance. To resolve this critical issue, we have developed a Fuzzy Outlier Removal (FOR) method that innovatively combines fuzzy set theory with information-theoretic metrics. The method calculates the informativeness of each point by constructing the membership of each point to the point cloud space in each dimension, and filters the points according to the order of the informativeness magnitude and the proportion of outlier points. FOR filters out a large amount of noise and outliers when filtering the point cloud to reduce detection time and improve efficiency, and retains the point cloud in the center region to reduce the loss of valid points and improve accuracy, which balances the contradiction between efficiency and accuracy.
The experimental results indicate that the performance of FOR is stable across different datasets and detection models. On the KITTI dataset, FOR achieves a competitive performance in reducing inference time while minimizing detection accuracy loss compared to other outlier removal methods. On the nuScenes dataset, FOR demonstrates a stable performance, though specific methods may excel in particular metrics (e.g., NiOR achieves higher NDS, while DROR achieves higher FPS). These mixed results suggest that FOR’s strength lies in its consistent ability to balance accuracy and efficiency, particularly in center-region preservation, rather than dominating all metrics across all datasets. Future work will focus on developing specific response mechanisms for local density variations, as the current method only considers the global position of points.

Author Contributions

Conceptualization, L.G. and Y.L.; methodology, L.G.; validation, L.G. and Y.L.; formal analysis, L.G.; writing—original draft preparation, L.G.; writing—review and editing, L.G.; visualization, Y.W.; supervision, X.A.; project administration, Z.Y.; funding acquisition, Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key Research and Development Program of China (2024YFC3014900).

Data Availability Statement

The KITTI dataset is publicly available at http://www.cvlibs.net/datasets/kitti/ (accessed on 29 April 2026), and the nuScenes dataset is publicly available at https://www.nuscenes.org/ (accessed on 29 April 2026). Both datasets are widely used benchmarks in the autonomous driving community. The implementation code of FOR and the experimental configurations will be made available from the corresponding author upon reasonable request. The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

The specific mAP values of Figure 4, Figure 5, Figure 6 and Figure 7 in the text are shown in Appendix A.
Table A1. The mAP% values of 3D bounding box (Figure 4).
Table A1. The mAP% values of 3D bounding box (Figure 4).
MetholdAP11AP40
MVFMVXNetPart-A2MVFMVXNetPart-A2
original data66.94 65.07 70.09 66.77 65.16 71.12
FOR-0.2563.02 62.82 66.73 63.04 62.78 67.48
ROR57.54 56.92 62.44 57.25 56.49 62.60
SOR54.11 53.92 57.55 53.15 53.05 57.62
DROR58.80 57.07 64.26 58.63 56.86 64.45
NiOR45.20 49.25 52.91 43.45 47.97 51.93
Table A2. The mAP% values of 2D bounding box (Figure 5).
Table A2. The mAP% values of 2D bounding box (Figure 5).
MetholdAP11AP40
MVFMVXNetPart-A2MVFMVXNetPart-A2
original data77.28 75.56 78.65 78.45 76.84 78.65
FOR-0.2574.10 73.41 75.95 75.56 74.65 77.72
ROR67.94 68.17 71.48 68.77 68.44 72.47
SOR64.83 64.86 68.22 64.45 65.26 68.32
DROR71.23 69.76 74.26 71.83 70.62 75.79
NiOR56.95 61.56 63.15 56.89 61.70 63.16
Table A3. The mAP% values of BEV (Figure 6).
Table A3. The mAP% values of BEV (Figure 6).
MetholdAP11AP40
MVFMVXNetPart-A2MVFMVXNetPart-A2
original data72.74 70.84 74.14 73.60 71.71 75.12
FOR-0.2569.53 68.89 71.06 69.96 69.15 71.71
ROR63.92 63.10 67.43 64.04 63.00 67.77
SOR60.96 59.91 63.06 60.48 59.63 62.87
DROR66.43 64.86 69.08 66.40 64.91 69.88
NiOR52.43 55.38 58.40 51.30 54.47 57.93
Table A4. The mAP% values of AOS (Figure 7).
Table A4. The mAP% values of AOS (Figure 7).
MetholdAP11AP40
MVFMVXNetPart-A2MVFMVXNetPart-A2
original data74.53 64.99 76.59 75.35 66.11 78.46
FOR-0.2571.26 62.90 74.10 72.35 63.92 75.62
ROR65.37 58.58 69.51 65.80 58.77 70.23
SOR62.34 55.26 66.06 61.64 55.49 65.87
DROR68.19 58.91 72.28 68.38 59.57 73.55
NiOR53.60 52.51 59.64 53.04 52.86 59.19

References

  1. Chen, S.; Liu, B.; Feng, C.; Vallespi-Gonzalez, C.; Wellington, C. 3d point cloud processing and learning for autonomous driving. arXiv 2020, arXiv:2003.00601. [Google Scholar] [CrossRef]
  2. Hitchcox, T.; Zhao, Y.F. Random walks for unorganized point cloud segmentation with application to aerospace repair. Procedia Manuf. 2018, 26, 1483–1491. [Google Scholar] [CrossRef]
  3. Qin, N.; Tan, W.; Guan, H.; Wang, L.; Ma, L.; Tao, P.; Fatholahi, S.; Hu, X.; Li, J. Towards intelligent ground filtering of large-scale topographic point clouds: A comprehensive survey. Int. J. Appl. Earth Obs. Geoinf. 2023, 125, 103566. [Google Scholar] [CrossRef]
  4. Chen, H.; Shen, J. Denoising of point cloud data for computer-aided design, engineering, and manufacturing. Eng. Comput. 2018, 34, 523–541. [Google Scholar] [CrossRef]
  5. Rong, Y.; Ma, L.; Zhou, Y. L0 Normal Optimization and Feature Rectification for Point Cloud Denoising. J. Comput.-Aided Des. Comput. Graph. 2024, 36, 1805–1815. [Google Scholar] [CrossRef]
  6. Lin, J.; Yin, H.; Yan, J.; Ge, W.; Zhang, H.; Rigoll, G. Improved 3D Object Detector Under Snowfall Weather Condition Based on LiDAR Point Cloud. IEEE Sens. J. 2022, 22, 16276–16292. [Google Scholar] [CrossRef]
  7. Charron, N.; Phillips, S.; Waslander, S.L. De-noising of lidar point clouds corrupted by snowfall. In Proceedings of the 2018 15th Conference on Computer and Robot Vision (CRV), Toronto, ON, Canada, 8–10 May 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 254–261. [Google Scholar]
  8. Han, X.F.; Jin, J.S.; Wang, M.J.; Jiang, W.; Gao, L.; Xiao, L. A review of algorithms for filtering the 3D point cloud. Signal Process. Image Commun. 2017, 57, 103–112. [Google Scholar] [CrossRef]
  9. Zhou, L.; Sun, G.; Li, Y.; Li, W.; Su, Z. Point cloud denoising review: From classical to deep learning-based approaches. Graph. Models 2022, 121, 101140. [Google Scholar] [CrossRef]
  10. Duan, Y.; Yang, C.; Li, H. Low-complexity adaptive radius outlier removal filter based on PCA for lidar point cloud denoising. Appl. Opt. 2021, 60, E1–E7. [Google Scholar] [CrossRef]
  11. Prio, M.H.; Patel, S.; Koley, G. Implementation of Dynamic Radius Outlier Removal (DROR) Algorithm on LiDAR Point Cloud Data with Arbitrary White Noise Addition. In Proceedings of the 2022 IEEE 95th Vehicular Technology Conference: (VTC2022-Spring), Helsinki, Finland, 19–22 June 2022; pp. 1–7. [Google Scholar] [CrossRef]
  12. Szutor, P.; Zichar, M. Fast Radius Outlier Filter Variant for Large Point Clouds. Data 2023, 8, 149. [Google Scholar] [CrossRef]
  13. Balta, H.; Velagic, J.; Bosschaerts, W.; De Cubber, G.; Siciliano, B. Fast statistical outlier removal based method for large 3D point clouds of outdoor environments. IFAC-PapersOnLine 2018, 51, 348–353. [Google Scholar] [CrossRef]
  14. Carrilho, A.C.; Galo, M.; Santos, R.C. Statistical outlier detection method for airborne LiDAR data. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2018, 42, 87–92. [Google Scholar] [CrossRef]
  15. Rakotosaona, M.J.; La Barbera, V.; Guerrero, P.; Mitra, N.J.; Ovsjanikov, M. Pointcleannet: Learning to denoise and remove outliers from dense point clouds. Comput. Graph. Forum 2020, 39, 185–203. [Google Scholar] [CrossRef]
  16. Guerrero, P.; Kleiman, Y.; Ovsjanikov, M.; Mitra, N.J. PCPNet: Learning local shape properties from raw point clouds. Comput. Graph. Forum 2018, 37, 75–85. [Google Scholar] [CrossRef]
  17. Pistilli, F.; Fracastoro, G.; Valsesia, D.; Magli, E. Learning Robust Graph-Convolutional Representations for Point Cloud Denoising. IEEE J. Sel. Top. Signal Process. 2021, 15, 402–414. [Google Scholar] [CrossRef]
  18. Li, Y.; Sheng, H. A single-stage point cloud cleaning network for outlier removal and denoising. Pattern Recognit. 2023, 138, 109366. [Google Scholar] [CrossRef]
  19. Kurup, A.; Bos, J. Dsor: A scalable statistical filter for removing falling snow from lidar point clouds in severe winter weather. arXiv 2021, arXiv:2109.07078. [Google Scholar] [CrossRef]
  20. Park, J.I.; Park, J.; Kim, K.S. Fast and accurate desnowing algorithm for LiDAR point clouds. IEEE Access 2020, 8, 160202–160212. [Google Scholar] [CrossRef]
  21. Zhao, Q.; Gao, X.; Li, J.; Luo, L. Optimization algorithm for point cloud quality enhancement based on statistical filtering. J. Sens. 2021, 2021, 7325600. [Google Scholar] [CrossRef]
  22. Roriz, R.; Campos, A.; Pinto, S.; Gomes, T. DIOR: A hardware-assisted weather denoising solution for LiDAR point clouds. IEEE Sens. J. 2021, 22, 1621–1628. [Google Scholar] [CrossRef]
  23. Li, X.; Lu, J.; Ding, H.; Sun, C.; Zhou, J.T.; Chee, Y.M. PointCVaR: Risk-Optimized Outlier Removal for Robust 3D Point Cloud Classification. Proc. AAAI Conf. Artif. Intell. 2024, 38, 21340–21348. [Google Scholar] [CrossRef]
  24. Rastiveis, H.; Shams, A.; Sarasua, W.A.; Li, J. Automated extraction of lane markings from mobile LiDAR point clouds based on fuzzy inference. ISPRS J. Photogramm. Remote Sens. 2020, 160, 149–166. [Google Scholar] [CrossRef]
  25. Zhang, C.; Zhou, H.; Chen, B.; Peng, Y.; Duan, J. Hybrid simplification algorithm for unorganized point cloud based on two-level fuzzy decision making. Optik 2023, 276, 170642. [Google Scholar] [CrossRef]
  26. Zhong, M.; Li, C.; Liu, L.; Wen, J.; Ma, J.; Yu, X. Fuzzy neighborhood learning for deep 3-D segmentation of point cloud. IEEE Trans. Fuzzy Syst. 2020, 28, 3181–3192. [Google Scholar] [CrossRef]
  27. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? the kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 3354–3361. [Google Scholar]
  28. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G. nuscenes: A multimodal dataset for autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11621–11631. [Google Scholar]
  29. Zhou, Y.; Sun, P.; Zhang, Y.; Anguelov, D.; Gao, J.; Ouyang, T.; Guo, J.; Ngiam, J.; Vasudevan, V. End-to-end multi-view fusion for 3d object detection in lidar point clouds. In Proceedings of the 3rd Conference on Robot Learning (CoRL), Osaka, Japan, 30 October–1 November 2019; pp. 923–932. [Google Scholar]
  30. Sindagi, V.A.; Zhou, Y.; Tuzel, O. Mvx-net: Multimodal voxelnet for 3d object detection. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 7276–7282. [Google Scholar]
  31. Shi, S.; Wang, Z.; Shi, J.; Wang, X.; Li, H. From points to parts: 3d object detection from point cloud with part-aware and part-aggregation network. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2647–2664. [Google Scholar] [CrossRef] [PubMed]
  32. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. Pointpillars: Fast encoders for object detection from point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12697–12705. [Google Scholar]
  33. Ning, X.; Li, F.; Tian, G.; Wang, Y. An efficient outlier removal method for scattered point cloud data. PLoS ONE 2018, 13, e0201280. [Google Scholar] [CrossRef]
Figure 1. Visualization of the KITTI dataset; outliers are more distributed in the far distance.
Figure 1. Visualization of the KITTI dataset; outliers are more distributed in the far distance.
Sensors 26 03070 g001
Figure 2. FOR framework.
Figure 2. FOR framework.
Sensors 26 03070 g002
Figure 3. Visualization of filtering results.
Figure 3. Visualization of filtering results.
Sensors 26 03070 g003
Figure 4. The mAP of 3D bounding box.
Figure 4. The mAP of 3D bounding box.
Sensors 26 03070 g004
Figure 5. The mAP of 2D bounding box.
Figure 5. The mAP of 2D bounding box.
Sensors 26 03070 g005
Figure 6. The mAP of BEV.
Figure 6. The mAP of BEV.
Sensors 26 03070 g006
Figure 7. The mAP of AOS.
Figure 7. The mAP of AOS.
Sensors 26 03070 g007
Table 1. Time complexity statistics of FOR.
Table 1. Time complexity statistics of FOR.
StepOperationTime Complexity
Calculate boundary values for each axisIterate over each coordinate value of each point and compute the boundary value3*O(N)
Calculate the membership of each pointIterate over each pointO(N)
Calculate informativeness at each pointIterate over each point3*O(N)
Sorting informativenessUsing the Quick Sort AlgorithmO(N*logN)
Filtering point cloudsIterate over the informativeness at each pointO(N)
Table 2. Configuration of Experimental Environment.
Table 2. Configuration of Experimental Environment.
Experimental EnvironmentConfiguration/Version Description
CPUIntel Core (TM) i9-12900k, 24 cores
GPUNVIDIA GeForce RTX 3090, 24 GB
Programming LanguagePython 3.8.15
Deep Learning FrameworkPytorch 1.9.0, CUDA 11.1
Table 3. Pointpillars in the nuScenes-trainval dataset.
Table 3. Pointpillars in the nuScenes-trainval dataset.
Assessment of IndicatorsmAPmATEmASEmAOEmAVEmAEENDS
original data0.29180.46980.35041.54300.37520.19120.4072
ROR0.26070.48900.35551.54620.37920.20250.3877
SOR0.27260.47960.35301.54710.36420.19660.3970
DROR0.25620.49590.35731.54800.39920.20380.3825
NiOR0.28200.47580.35201.54640.37400.19460.4014
FOR-0.250.27200.47990.35211.54700.37020.19650.3961
Table 4. FPS comparison of different models on different datasets.
Table 4. FPS comparison of different models on different datasets.
Model
(Dataset)
MVF
(KITTI)
MVXNet (KITTI)Part-A2
(KITTI)
Pointpillars
(nuScenes)
original data24.1511.048.0921.16
FOR-0.2533.8214.9318.1527.00
ROR23.5712.098.8123.23
SOR24.3311.818.6121.73
DROR23.2412.338.5534.18
NiOR25.1312.549.0322.35
Table 5. Parameter sensitivity comparison of FOR.
Table 5. Parameter sensitivity comparison of FOR.
NormThresholdMVFMVXNetPart-A2
3DFOR-0.166.3464.7169.94
FOR-0.1565.6064.4569.21
FOR-0.264.6363.5768.45
2DFOR-0.177.0575.4178.94
FOR-0.1576.6775.1078.51
FOR-0.275.7574.7377.93
BEVFOR-0.172.5370.8873.94
FOR-0.1571.6070.5673.59
FOR-0.270.9269.6472.64
AOSFOR-0.174.0864.9976.74
FOR-0.1573.7264.4876.33
FOR-0.272.8163.5275.79
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gan, L.; Yang, Z.; Liu, Y.; Wang, Y.; An, X. FOR: Point Cloud Outlier Removal Based on Fuzzy Theory and Informativeness and Its Application to 3D Object Detection. Sensors 2026, 26, 3070. https://doi.org/10.3390/s26103070

AMA Style

Gan L, Yang Z, Liu Y, Wang Y, An X. FOR: Point Cloud Outlier Removal Based on Fuzzy Theory and Informativeness and Its Application to 3D Object Detection. Sensors. 2026; 26(10):3070. https://doi.org/10.3390/s26103070

Chicago/Turabian Style

Gan, Lili, Zhengyi Yang, Yiyi Liu, Yaqi Wang, and Xinyan An. 2026. "FOR: Point Cloud Outlier Removal Based on Fuzzy Theory and Informativeness and Its Application to 3D Object Detection" Sensors 26, no. 10: 3070. https://doi.org/10.3390/s26103070

APA Style

Gan, L., Yang, Z., Liu, Y., Wang, Y., & An, X. (2026). FOR: Point Cloud Outlier Removal Based on Fuzzy Theory and Informativeness and Its Application to 3D Object Detection. Sensors, 26(10), 3070. https://doi.org/10.3390/s26103070

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop