Next Article in Journal
Using Vector Databases for the Selection of Related Occupations: An Empirical Evaluation Using O*NET
Previous Article in Journal
Boost-Classifier-Driven Fault Prediction Across Heterogeneous Open-Source Repositories
Previous Article in Special Issue
Self-Supervised Foundation Model for Template Matching
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RecurrentOcc: An Efficient Real-Time Occupancy Prediction Model with Memory Mechanism

by
Zimo Chen
,
Yuxiang Xie
and
Yingmei Wei
*
Laboratory for Big Data and Decision, National University of Defense Technology, Changsha 410003, China
*
Author to whom correspondence should be addressed.
Big Data Cogn. Comput. 2025, 9(7), 176; https://doi.org/10.3390/bdcc9070176
Submission received: 21 April 2025 / Revised: 7 June 2025 / Accepted: 30 June 2025 / Published: 2 July 2025
(This article belongs to the Special Issue Perception and Detection of Intelligent Vision)

Abstract

Three-dimensional Occupancy Prediction provides a detailed representation of the surrounding environment, essential for autonomous driving. Long temporal image sequence fusion is a common technique used to improve the occupancy prediction performance. However, existing temporal fusion methods are inefficient due to three issues: repetitive feature extraction from temporal images, redundant fusion of temporal features, and suboptimal fusion of long-term historical features. To address these challenges, we propose the Recurrent Occupancy Prediction Network (RecurrentOcc). We introduce the Scene Memory Gate, a new temporal fusion module that condenses temporal scene features into a single historical feature map. This eliminates the need for repeated extraction and aggregation of multiple temporal images, reducing computational overhead. The Scene Memory Gate selectively retains valuable information from historical features and recurrently updates the historical feature map, enhancing temporal fusion performance. Additionally, we design a simple yet efficient encoder, significantly reducing the number of model parameters. Compared with other real-time methods, RecurrentOcc achieves state-of-the-art performance of 39.9 mIoU on the Occ3D-NuScenes dataset with the fewest parameters of 59.1 M and an inference speed of 23.4 FPS.

1. Introduction

In autonomous driving, the ability to perceive and understand the surrounding environment is a fundamental requirement. 3D occupancy prediction, which reconstructs a fine-grained representation of the street scene in the form of 3D voxels, has gained significant attention in recent years. Although previous works [1,2,3,4] have significantly improved the performance of 3D occupancy prediction, they have also resulted in a corresponding increase in model scale and computation. Due to the limited computational resources available on vehicles, 3D occupancy prediction models should have a compact model size and low computational demands for practical deployment. Additionally, to ensure safety, the model must achieve real-time inference speed. Therefore, some researchers focus on improving the model’s computational efficiency, such as SparseOcc [5], FlashOcc [6], and FastOcc [7].
To enhance occupancy prediction performance, many methods [5,6,8,9,10] have incorporated temporal image sequences. Historical images provide supplementary information, aiding the model in reconstructing the current scene, especially in occluded areas. However, integrating temporal images, particularly long-term sequences, increases computational costs for image processing and feature aggregation, thus reducing the model’s inference speed. Figure 1a illustrates a widely used method proposed by [11] for temporal feature aggregation. It reuses the backbone to extract features from historical images and feeds these temporal features into a fusion module for aggregation. The fusion module typically consists of several convolutional layers. However, this approach has three main drawbacks, leading to computational inefficiency. First, the extraction of historical image features introduces redundant computation, since these image features have already been extracted at previous timestamps. Second, the temporal fusion process is redundant. For example, let T represent the current time and K represent the sequence length. At T 1 , the module aggregates features from [ T 1 K , T 1 ] , whereas at T, it aggregates features from [ T K , T ] , causing redundant fusion of features in the overlapping interval [ T K , T 1 ] . Third, current temporal fusion methods risk introducing noise. These methods directly concatenate short-term and long-term temporal features with current features and aggregate them indiscriminately. While long-term features may provide valuable information about occluded areas (inaccessible to short-term observations), they often contain outdated scene information due to the ego-vehicle’s motion. Direct fusion without proper filtering may therefore propagate irrelevant or outdated information, adversely affecting model performance. GSD-Occ [12] introduces a temporal feature queue to store the computed scene features, as shown in Figure 1b. When reconstructing the current scene, the model simply retrieves the adjacent historical features from the queue, avoiding redundant feature extraction [3,6,9,11]. Nevertheless, GSD-Occ’s temporal fusion still exhibits the last two drawbacks. Moreover, the size of stored features grows with the length of the temporal image sequence.
In this paper, we propose a new paradigm for temporal fusion, as shown in Figure 1c. The network condenses the historical information into a single feature (we refer to it as a historical feature map) and recurrently updates this feature map over time. As a result, the network integrates only the historical feature map with the current scene feature for temporal fusion, avoiding the redundant computation caused by temporal image sequence fusion. We introduce the Scene Memory Gate (SMG) as the temporal fusion module, which selectively integrates the historical feature map with the current scene feature and retains the fused feature for the next iteration, mitigating the third challenge. Furthermore, we design the Light Encoder, a simpler and more efficient encoder structure that adopts the VRWKV [13] network as its basic block. The Light Encoder significantly reduces the number of model parameters while maintaining high inference speed and prediction performance. In summary, our contributions are as follows:
  • We introduce a new temporal fusion module, SMG, which selectively condenses valuable temporal information into a historical feature map. It allows the model to integrate only a single feature map instead of multiple temporal features, improving computational efficiency and achieving superior performance.
  • We design a simpler encoder structure based on the VRWKV network, termed the Light Encoder. The introduction of this encoder reduces the model’s parameters to only 59.1 M without compromising performance.
  • Our proposed Recurrent Occupancy Prediction Network (RecurrentOcc, ReOcc) achieves state-of-the-art performance of 39.9 mIoU on the Occ3D-NuScenes dataset [14] with the fewest parameters and an inference speed of 23.4 FPS, compared with other real-time competitors [5,6,9,12].

2. Related Works

2.1. Vision-Based 3D Occupancy Prediction

Monoscene [1] is the first vision-based occupancy prediction work that reconstructs scenes solely from RGB images. It proposes FLoSP to transform 2D image features into 3D features and uses a 3D U-Net architecture to predict voxel occupancy and semantic classification. VoxFormer [2] introduces a two-stage framework, where sparse voxels are identified in the first stage and the scene is refined in the second. OccFormer [15] builds a transformer-based encoder-decoder network. It utilizes a dual-path transformer [16] block as an encoder to extract the global and local features of the scene and Mask2Former [17] as a decoder for semantic occupancy prediction. SurroundOcc [18] employs deformable attention [19] to directly create a dense 3D representation from multi-camera images without first predicting depth. Instead of voxel-wise prediction, Symphonies [20] creates instance queries and reconstructs the scene using instance semantics, improving contextual understanding. COTR [3] proposes a geometry-aware encoder, consisting of explicit and implicit view transformation, to generate a compact 3D representation. In recent years, numerous advanced methods have emerged, rapidly improving occupancy prediction performance. However, many of these methods overlook deployment challenges and suffer from high computational costs. Consequently, researchers have increasingly focused on developing lightweight models suitable for vehicle deployment.

2.2. Deployment-Friendly Occupancy Prediction

While dense 3D voxels provide an effective representation for autonomous driving, their computation is time-consuming and memory-intensive. To simplify intermediate computations, TPVFormer [21] reduces the 3D space to tri-perspective views, and FlashOcc [6] uses Bird’s Eye View (BEV) for scene feature representation. Given that more than half of the voxels in a scene are empty, SparseOcc [5] employs a sparse transformer to minimize computational costs. OctreeOcc [22] uses an octree-based [23] architecture to construct the occupancy prediction network, significantly reducing the number of voxels needed to depict the scene. GSD-Occ [12] introduces a dual-branch network, with a 2D encoder for BEV semantic perception and a 3D encoder leveraging the re-parameterization technique [24] for 3D geometry perception. Despite significant progress in deployment-friendly occupancy prediction, we find that the temporal fusion process is still inefficient. This inspires us to seek a more efficient temporal fusion approach.

2.3. Temporal Fusion in 3D Occupancy Prediction

Incorporating temporal frames into the input is a common practice to improve the prediction performance. A widely used approach is to concatenate temporal features extracted from the backbone and aggregate them using convolutional operations [11]. This approach overlooks the varying contributions of features across different times and locations, potentially leading to insufficient aggregation. Moreover, the repeated feature extraction operation introduces redundant computation. HTCL [25] proposes a new paradigm for improving temporal feature aggregation, which includes cross-frame affinity measurement and affinity-based dynamic refinement. While this method greatly improves prediction performance, it also increases computational complexity. Inspired by [26], GSD-Occ [12] employs a history feature queue to store temporal features, avoiding redundant extraction, but it still relies on long-term temporal fusion to achieve better results. In this paper, we propose the scene memory gate that aggregates temporal features from historical frames into a single feature map and continuously updates it over time, enabling efficient temporal fusion and real-time inference.

3. Method

3.1. Problem Setup

The task input is a sequence of multi-view temporal images I R b × t × n × 3 × h × w , where b denotes the batch size, t denotes the temporal sequence length, n denotes the number of camera views, and h , w denote the image resolution. The output of the occupancy prediction model G is a 3D representation of the scene with semantic labels O R H × W × Z × C , where H , W , Z denote the spatial resolution of the scene, and C denotes the number of labels. The labels consist of C 1 object categories and an ’empty’ category. If a voxel in the scene is occupied, G assigns it an object category; otherwise, it assigns the ’empty’ category.

3.2. Overall Architecture

Figure 2 illustrates the overall architecture of ReOcc. First, the input images I at time T are processed by the backbone to extract image features F R b × n × c × h × w , where c denotes the feature dimension. Features F are then lifted to 3D feature V T R b × C × H 0 × W 0 × Z 0 in the 2D-to-3D transformation module, where C and H 0 , W 0 , Z 0 denotes V T ’s feature and spatial dimensions, respectively. Traditional occupancy networks employ 3D encoders to aggregate the semantic and geometry features from V T , which is computationally expensive. To reduce complexity, we adopt FlashOcc’s [6] approach by transforming the 3D features into Bird’s Eye View (BEV) representations. Although this projection sacrifices height-dimension information, it achieves a trade-off for faster inference. Critically, the compressed 2D features retain implicit height cues, which can assist the network to reconstruct 3D features from the BEV space.
Building upon this, we adopt the dual-path architecture from GSD-Occ [12], which separately processes semantic information in 2D and geometric information in 3D. The semantic path incorporates a Scene Memory Gate for efficient temporal fusion along with a lightweight 2D encoder, while the geometric path maintains GSD-Occ’s original design, which consists of a non-dilated small-kernel 3D convolution and multiple dilated small-kernel 3D convolutions. During inference, these small-kernel convolutions are re-parameterized into a large-kernel convolution to accelerate calculation. Specifically, the BEV feature map B T R b × C × H 0 × W 0 is generated by averaging V T ’s features along the height dimension. The Scene Memory Gate aggregates B T with the historical scene feature map h T 1 and outputs the temporal fused features h T , which are then saved as the next historical scene feature map. Simultaneously, h T is restored to the 3D feature V b via the BEV-Voxel Lifting (BVL) module [12]. V b is combined with V T to generate V T . Next, the 3D geometric feature V T and the 2D semantic feature h T are further processed by 3D and 2D encoders, respectively, producing V g and B s . B s is then restored to 3D semantic feature V s via the BVL method. Finally, V s and V g are combined, and the resulting feature is processed by a prediction head to obtain the final prediction O. The dual-path architecture enables computationally and memory-efficient processing of dense, high-dimensional semantic features through its 2D pathway, while the 3D geometric path preserves crucial spatial information that compensates for height dimension loss in the semantic processing.
Two-dimensional-to-three-dimensional transformation. Two-dimensional-to-three-dimensional transformation aggregates multi-view image features to generate the corresponding 3D scene feature. We adopt the LSS method [27], where image features F R b × n × c × h × w are fed into a depth net and a context net [28] to predict the depth distribution D R b × n × d × h × w and semantic feature F R b × n × c × h × w , respectively, where d denotes the number of bins. The outer product of F and D generates a pseudo 3D point cloud feature P R b × n × c × d × h × w , which is then converted into 3D voxel features V R b × c × H 0 × W 0 × Z 0 via voxel-pooling based on the camera intrinsic parameters K i i = 1 n and extrinsic parameters [ R i | t i ] i = 1 n .
BEV-to-3D transformation. BEV-to-3D transformation aims to recover the height information from the BEV feature. We use the BVL method [12], which is illustrated in the top right of Figure 2. The approach of BVL shares similarities with LSS. It consists of two branches: one height branch employs a convolutional network to predict height distribution based on BEV feature information, while the other context branch generates height-aware features via a separate convolutional network. Subsequently, the 3D voxel features are obtained through the outer product of the height distribution and height-aware features. For example, taking B s R b × C × H 0 × W 0 as input, BVL converts B s into height distribution H s R b × Z 0 × H 0 × W 0 and height-aware features B s R b × C × H 0 × W 0 . Then B s and H s are computed by the outer product B s H s to generate the 3D feature V s R b × C × H 0 × W 0 × Z 0 .

3.3. Scene Memory Gate

Inspired by LSTM [29], SMG employs a recurrent process for temporal fusion. The core idea of SMG is to selectively retain valuable information from historical scene features and discard features from relatively distant or highly uncertain areas. The retained historical features are then integrated into the current BEV feature. The fused feature is then saved as the historical feature map, preparing for the next iteration of temporal fusion. Due to pose changes, the historical feature map needs to be reprojected onto the current BEV plane before fusion. The detailed structure of SMG is illustrated in Figure 3.
SMG first transforms the historical scene feature h T 1 into a hidden state s T 1 through a hidden block comprising a 3 × 3 convolution layer and a point-wise convolution layer, with each layer succeeded by a normalization and ReLU activation layer. Next, a separate 3 × 3 convolution layer followed by a sigmoid activation layer is used to predict the value score of each pixel in h T 1 based on s T 1 . The same module is also applied to compute the score of the current BEV feature B T . The mask M T 1 is derived from Equation (1):
M T 1 = σ f s ( s T 1 ) + ( 1 σ ( f c ( B T ) ) )
where f c ( · ) and f s ( · ) denote convolution operations, and σ ( · ) denotes the sigmoid function. The term 1 σ ( f c ( B T ) ) is used to enhance the weight of historical features in highly uncertain areas, such as occluded areas, of the current scene. Due to the limitations of the LLS method [27], the initial BEV feature B is sparse, as it cannot infer features from occluded regions. The introduction of 1 σ ( f c ( B T ) ) enhances the retention of historical features at corresponding positions. Furthermore, the mask values are bounded within the interval [0,1], as formulated in Equation (2).
M T 1 ( x ) = M T 1 ( x ) if M T 1 ( x ) 1 1 if M T 1 ( x ) > 1
The selected valuable historical feature is derived from Equations (2) and (3) and generates the fused feature map:
h ˜ T 1 = M T 1 h T 1
h T = h ˜ T 1 + B T
For the first BEV feature map B 1 in an image sequence (where no historical features exist), we initialize the hidden state as
h 0 = B 1
Compared with the temporal feature queue [12], SMG is more memory friendly as it only needs to save a single feature map rather than multiple. And its mask mechanism reduces the risk of noise contamination in the feature fusion process. Additionally, because SMG is a recurrent operation over time, there exists long-term memory forgetting. We argue that this mechanism is well-suited for autonomous driving. In most cases, autonomous vehicles move forward, making the information from earlier scenes less relevant for the current scene reconstruction. Long-term forgetting can automatically discard features from earlier scenes, allowing the network to focus more on the current moment and recent temporal features. Therefore, our designed SMG can achieve excellent temporal feature aggregation performance by fusing only one historical feature map.

3.4. Light Encoder

While SMG achieves efficient temporal feature fusion, it introduces additional computational overhead and parameters compared with convolution-based methods when fusing two temporal features. Therefore, we design a simpler and more efficient 2D encoder to work alongside the SMG to balance computational speed, network size, and performance. The Light Encoder’s architecture is illustrated on the left of Figure 4. The encoder consists of two encoding blocks. A downsample layer is inserted to obtain high-dimensional semantic features. The high-dimensional feature and low-dimensional feature are merged via a single convolution layer.
We aim to seek an encoded block structure that possesses both strong reasoning capabilities and fast inference speeds. Therefore, we employ VRWKV [13] as the encoding block. The architecture of VRWKV is illustrated on the right side of Figure 4. The input is first fed into the spatial mix module. After the Q-shift operation, the input is transformed to r s , k s , and v s . Then, the global attention result w k v is obtained through the bidirectional attention mechanism Bi-WKV. Bi-WKV has linear complexity and can be expressed in a summation form:
w k v t = Bi WKV ( K , V ) t = i = 0 , i t L 1 e ( | t i | 1 ) / L · w + k i v i + e u + k t v t i = 0 , i t L 1 e ( | t i | 1 ) / L · w + k i + e u + k t .
where L denotes the length of the sequence. w k v is multiplied by σ ( r s ) and passed through a linear projection to generate the output of the spatial mix module O s :
O s = F l ( σ ( r s ) w k v )
where σ ( · ) denotes the sigmoid function, ⊙ denotes the Hadamard product, and F l denotes the linear projection. The channel mix module follows a similar calculation, formulated as
v c = F l ( S q u a r e d R e L U ( k c ) )
O c = F l ( σ ( r c ) v c )
For comparison, we also implement VMamba [30] and ResNet [31] as basic encoding blocks. Finally, we chose to employ the VRWKV network. The detailed experiments will be elaborated on in Section 4.5.

3.5. Loss Function

Adhering to common practices [1,6,10,12], the loss function consists of four parts: the binary CE loss L g e o , CE loss L s e m , LovaszSoftmax loss L L o v a s z , and focal loss L f o c a l .
L t o t a l = L g e o + L s e m + L l o v a s z + L f o c a l

4. Experiment

4.1. Datasets

We conduct experiments on the NuScenes dataset [32], a comprehensive autonomous driving dataset consisting of 1000 driving scenarios sourced from Boston and Singapore. Each scenario spans 20 s, totaling approximately 15 h of driving data. These data cover a wide range of locations, times, and weather conditions. The NuScenes dataset offers abundant sensor data, including six cameras, one LiDAR (Light Detection and Ranging) sensor, five radars, and GPS and IMU (Inertial Measurement Unit) sensor data. The occupancy ground truth is provided by Occ3d-NuScenes [14]. It includes 700 scenes for training and 150 scenes for validation. Each scene covers a range of [−40 m, −40 m, −1 m, 40 m, 40 m, 5.4 m] with a voxel size of [0.4 m, 0.4 m, 0.4 m]. The labels include 16 known object categories, an ‘others’ class for unknown objects, and an `empty’ class for unoccupied voxels.

4.2. Evaluation Metrics

We report the mean Intersection over Union (mIoU) and Ray-level mIoU (RayIoU) [5] to evaluate the prediction performance of our model. We also evaluate the model’s number of parameters, frames per second (FPS), and storage size of temporal data.

4.3. Implementation Details

We adopt ResNet-50 [31] as the image backbone, and the input image resolution is set to 704 × 256 . We propose two versions of our model. The feature dimensions of ReOcc-B and ReOcc-S are set to 128 and 64, respectively. Our model is trained on 8 NVIDIA 4090 GPUs for 24 epochs with a batch size of 3. The optimizer is Adam [33] with a learning rate of 1 × 10 4 and a weight decay of 0.01 . Following previous research [5,9,12], the inference speed is tested on a single NVIDIA A100 GPU with a batch size of 1. Historical features are saved using the ’savez_compressed’ function from the NumPy package to measure their storage size.

4.4. Main Results

We compare ReOcc with recent state-of-the-art (SOTA) methods on Occ3D-NuScenes in Table 1. When trained with a visible mask, ReOcc-B achieves 39.9 mIoU and 31.5 RayIoU on the NuScenes validation dataset with an inference speed of 23.4 FPS. Without the mask, ReOcc-B achieves 31.9 mIoU and 38.2 RayIoU. ReOcc-S, with reduced feature dimensions, shows a slight drop of 1.7 % mIoU and 0.3 % RayIoU with the mask and 0.9 % mIoU and 1.0 % RayIoU without the mask but improves FPS by 12.8 % over ReOcc-B.
To demonstrate the efficiency of our model, we perform a more detailed comparison with other real-time methods, as shown in Table 2. Fusing only a single feature map, our model achieves comparable performance to GSD-Occ(16f) [12] and Panoptic-FlashOcc(8f) [9] and outperforms SparseOcc(8f) [5] by 1.8 mIoU and 4.2 RayIoU. We provide a qualitative comparison with GSD-Occ [12] in Figure 5. Our method demonstrates stable reconstruction and superior perception capabilities. In terms of inference speed, ReOcc-B and ReOcc-S outperform SparseOcc(8f) [5] by 6.1 FPS and 9.1 FPS, GSD-Occ(16f) [12] by 3.4 FPS and 6.4 FPS, and GSD-Occ(2f) [12] by 2.4 FPS and 5.4 FPS, respectively. However, they are still around 10 FPS slower than the Panoptic-FlashOcc series. Although Panoptic-FlashOcc [9] boasts very fast inference speed, it still uses the temporal fusion structure shown in Figure 1a. The inference speed test excludes the time required to extract features from historical images. Therefore, Panoptic-FlashOcc [9] cannot avoid redundant calculations in the temporal fusion process during the training stage. We include the time for extracting historical features and retest its inference speed. As shown by Panoptic-FlashOcc(2f)* and Panoptic-FlashOcc(8f)*, there’s a significant decrease in FPS, indicating that the model’s training is not efficient. In contrast, our model maintains a high inference speed during training, thus saving more training resources. Our model has the smallest number of parameters while maintaining promising performance and inference speed. The parameter count of ReOcc-S is 30.5 % lower than that of SparseOcc(8f) [5], 51.4 % lower than GSD-Occ(16f) [12], and 27.3 % lower than Panoptic-FlashOcc(8f) [9]. Additionally, since both GSD-Occ [12] and Panoptic-FlashOcc [9] utilize long-term temporal feature fusion, they need to store multiple historical features. We measure their sizes and find that the storage size of GSD-Occ(16f) [12] is 32.3 M, and that of Panoptic-FlashOcc [9] is 55.0 M. Thanks to the advantages of the SMG structure, ReOcc only needs to store a single feature map, with a size of approximately one-tenth of theirs. The storage size of ReOcc-S is only 1.9 M, even smaller than that of GSD-Occ(2f) [12] and Panoptic-FlashOcc(2f) [9]. We report ReOcc’s IoU for specific categories and provide additional quantitative results in Appendix A.

4.5. Ablation Studies

We conduct ablation experiments with ReOcc-B on the NuScenes dataset. All of the experiments are trained with a visible mask.
  • Ablation on different components: Table 3 demonstrates the effectiveness of the Spatial-Motion Gating (SMG) and Light Encoder. The baseline uses the GSD-Occ structure with a temporal fusion of 2 frames. Replacing the baseline’s temporal fusion method with SMG results leads to only minor changes in the model’s parameters and inference speed while increasing IoU by 3.5 % , mIoU by 7.0 % , and RayIoU by 6.0%. This demonstrates SMG’s effectiveness in preserving fused historical features. Replacing the baseline’s 2D encoder with the Light Encoder keeps mIoU and RayIoU nearly unchanged while reducing the parameter count by 16.8%, demonstrating the Light Encoder’s efficiency. When both modules are combined, the model’s inference speed slightly decreases compared with the baseline, but its IoU, mIoU, and RayIoU improve by 5.1%, 8.4%, and 6.3%, respectively, and its parameter count decreases by 16.5%.
    Ablation on Light Encoder: To explore a more efficient network structure for the encoded block, we conduct experiments with the VRWKV [13], VMamba [30], and ResNet [31] networks, as shown in Table 4. Compared with ResNet, VMamba improves IoU by 0.37, mIoU by 0.20, slightly decreases RayIoU by 0.03, reduces parameters by 0.8 M, and lowers inference speed by 1.5 FPS. VRWKV, on the other hand, achieves a higher IoU and mIoU by 0.43 and 0.28 but a lower RayIoU by 0.3. It has 0.4M fewer parameters but has a lower inference speed by 0.8 FPS. Considering mIoU as the primary metric, we choose the VRWKV network as the basic block in the Light Encoder. We also perform ablation experiments on the number of encode blocks. Increasing the number of blocks from 1 to 2 improves IoU by 0.21, mIoU by 0.3, and RayIoU by 0.03, but FPS decreases by 0.8, and the parameter count increases by 2.34 M. Increasing the number of blocks from 2 to 3 improves IoU by 0.37, mIoU by 0.23, and RayIoU by 0.12, but FPS decreases by 1.4, and the parameter count increases by 5.77 M. To balance the model’s performance, size, and inference speed, we set the number of blocks to 2 in this paper.
    Ablation on SMG: To further validate the effectiveness and efficiency of SMG on temporal fusion, we compared it with GSD-Occ’s temporal fusion method (denoted as GSD fusion) with 2-frame and 16-frame inputs, as shown in Table 5. SMG outperforms GSD fusion(16f) by 0.63 in IoU, 0.89 in mIoU, 0.41 in RayIoU, and 0.4 in FPS, while requiring only 3.53M of storage space for temporal features. Although GSD fusion(2f) has fewer parameters and smaller storage compared with SMG, its performance is significantly inferior to SMG. SMG outperforms GSD fusion(2f) by 3.4% in IoU, 7.9% in mIoU, and 7.3% in RayIoU. This demonstrates SMG’s superior temporal feature fusion efficiency compared with GSD-Occ’s method. Moreover, to further validate the effectiveness of SMG’s mask mechanism, we conducted an ablation study by removing the mask module. As shown in Table 6, this modification resulted in performance degradation of 0.42 mIoU and 0.45 RayIoU. The result demonstrates that SMG’s mask mechanism effectively filters temporal feature information and enhances the model’s focus on semantically valuable regions. Additionally, we conducted an experiment to investigate whether SMG requires an active forgetting mechanism similar to LSTM’s forget gate, which is also present in Table 6. We implemented a simple forget gate (FG) consisting of a convolutional layer f g with kernel size 3 and a sigmoid activation layer. Before storing the history feature map, h T is first processed through this forget gate:
    h T = σ ( f g ( h T ) ) h T
    The addition of the forget gate yielded marginal IoU improvement while causing slight degradation in mIoU and RayIoU metrics. These results demonstrate that the active forgetting mechanism is unnecessary, as our existing mask mechanism already provides adequate filtering capability, and the recurrent computation inherently possesses long-term forgetting capacity.
    Ablation on 3D-to-BEV transformation: We compared two existing methods for compressing 3D features into 2D representations. The first approach performs mean pooling along the height dimension, while the second method, proposed by FlashOcc [6], initially merges both the feature dimension and height dimension into a single dimension and then applies a convolutional layer for channel compression. As shown in Table 7, we denote these methods as ’mean pooling’ and ’convolution’, respectively. Both methods achieve comparable IoU scores, with mean pooling showing a slight advantage of +0.33 mIoU and +0.21 RayIoU.

5. Conclusions

In this paper, we introduce an efficient occupancy prediction model named ReOcc. It uses an SMG module to aggregate and store historical feature information in a single historical feature map. This approach reduces redundant computations in temporal fusion and conserves space for storing feature data. Furthermore, we design a Light Encoder to further decrease the model’s parameter count and computational load. On the Occ3D-NuScenes validation dataset, our model achieves performance comparable to other real-time competitors that fuse multiple historical features, but with fewer parameters and under the condition of fusing only one historical feature map. This demonstrates the high efficiency of our model. While converting 3D features to BEV representation significantly improves model inference speed, the inherent loss of height-dimensional information in BEV features limits 3D reconstruction quality, even with supplementary 3D geometric features. This limitation becomes particularly apparent in cases where objects overlap vertically (e.g., a person standing under a tree), as BEV representations struggle to distinguish such conditions. Moreover, although our proposed Scene Memory Gate reduces redundant computation and storage compared with previous methods, its speed advantage remains modest. Future research will focus on developing more efficient temporal fusion mechanisms to accelerate inference, as well as exploring dual-view 2D image representations as potential alternatives to 3D features, thereby addressing BEV’s information loss.

Author Contributions

Conceptualization, Y.X. and Z.C.; methodology, Z.C.; validation, Z.C.; formal analysis, Z.C.; investigation, Z.C. and Y.W.; writing—original draft preparation, Z.C.; writing—review and editing, Z.C. and Y.W.; visualization, Z.C.; supervision, Y.X. and Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Hunan Province, China, under Grant 2023JJ30082.

Data Availability Statement

The NuScenes dataset is available on https://www.nuscenes.org/ (accessed on 21 April 2025). The Annotation is available on https://github.com/Tsinghua-MARS-Lab/Occ3D (accessed on 21 April 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

In Table A1, we demonstrate the recognition and reconstruction capabilities of ReOcc-B and ReOcc-S for specific categories. In Figure A1, we visualize additional prediction results based on ReOcc-B. We present these results from a BEV perspective to facilitate a comprehensive display of the reconstructed scene.
Figure A1. Qualitative results on Occ3D-NuScenes validation set.
Figure A1. Qualitative results on Occ3D-NuScenes validation set.
Bdcc 09 00176 g0a1
Table A1. ReOcc’s 3D Occupancy prediction performance for specific categories on the Occ3D-NuScenes.
Table A1. ReOcc’s 3D Occupancy prediction performance for specific categories on the Occ3D-NuScenes.
MethodVis.
Mask
SC
IoU
SSC
mIoU
Bdcc 09 00176 i001 OthersBdcc 09 00176 i002 BarrierBdcc 09 00176 i003 BicycleBdcc 09 00176 i004 BusBdcc 09 00176 i005 CarBdcc 09 00176 i006 Const. Veh.Bdcc 09 00176 i007 MotorcycleBdcc 09 00176 i008 PedestrianBdcc 09 00176 i009 Traffic ConeBdcc 09 00176 i010 TrailerBdcc 09 00176 i011 TruckBdcc 09 00176 i012 Drive. Suf.Bdcc 09 00176 i013 Other FlatBdcc 09 00176 i014 SidewalkBdcc 09 00176 i015 TerrainBdcc 09 00176 i016 ManmadeBdcc 09 00176 i017 Vegetation
ReOcc-B70.5039.9712.3747.8028.7643.8251.2327.3228.0228.9429.4032.9437.6481.3443.5050.9655.5342.8637.01
ReOcc-S70.0239.2812.7847.2227.5444.3150.9026.3827.5828.2028.2331.0036.9280.7742.2150.3955.1041.7336.48
ReOcc-B45.9131.9811.2043.6424.7540.4243.9122.5126.8925.7427.9022.3632.4865.4036.7539.7536.7021.2622.01
ReOcc-S45.4731.6110.8742.6025.9339.5043.6222.3026.4625.6327.3521.4831.9565.0735.9139.0836.2221.0322.11

References

  1. Cao, A.-Q.; de Charette, R. Monoscene: Monocular 3d Semantic Scene Completion. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  2. Li, Y.; Yu, Z.; Choy, C.B.; Xiao, C.; Álvarez, J.M.; Fidler, S.; Feng, C.; Anandkumar, A. VoxFormer: Sparse Voxel Transformer for Camera-Based 3D Semantic Scene Completion. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023. [Google Scholar]
  3. Ma, Q.; Tan, X.; Qu, Y.; Ma, L.; Zhang, Z.; Xie, Y. COTR: Compact Occupancy TRansformer for Vision-Based 3D Occupancy Prediction. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  4. Wang, Y.; Chen, Y.; Liao, X.; Fan, L.; Zhang, Z. PanoOcc: Unified Occupancy Representation for Camera-based 3D Panoptic Segmentation. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  5. Liu, H.; Chen, Y.; Wang, H.; Yang, Z.; Li, T.; Zeng, J.; Chen, L.; Li, H.; Wang, L. Fully Sparse 3D Occupancy Prediction. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, Netherlands, 8–16 October 2023. [Google Scholar]
  6. Yu, Z.; Shu, C.; Deng, J.; Lu, K.; Liu, Z.; Yu, J.; Yang, D.; Li, H.; Chen, Y. FlashOcc: Fast and Memory-Efficient Occupancy Prediction via Channel-to-Height Plugin. arXiv 2023, arXiv:2311.12058. [Google Scholar]
  7. Hou, J.; Li, X.; Guan, W.; Zhang, G.; Feng, D.; Du, Y.; Xue, X.; Pu, J. FastOcc: Accelerating 3D Occupancy Prediction by Fusing the 2D Bird’s-Eye View and Perspective View. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024. [Google Scholar]
  8. Pan, M.; Liu, J.; Zhang, R.; Huang, P.; Li, X.; Liu, L.; Zhang, S. Renderocc: Vision-Centric 3d Occupancy Prediction with 2d Rendering Supervision. In Proceedings of the 2024 IEEE International Conference on Robotics and Automation (ICRA), Yokohama, Japan, 13–17 May 2024. [Google Scholar]
  9. Yu, Z.; Shu, C.; Sun, Q.; Linghu, J.; Wei, X.; Yu, J.; Liu, Z.; Yang, D.; Li, H.; Chen, Y. Panoptic-FlashOcc: An Efficient Baseline to Marry Semantic Occupancy with Panoptic via Instance Center. arXiv 2024, arXiv:2406.10527. [Google Scholar]
  10. Li, Z.; Yu, Z.; Wang, W.; Anandkumar, A.; Lu, T.; Alvarez, J.M. Fb-bev: Bev representation from forward-backward view transformations. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023. [Google Scholar]
  11. Huang, J.; Huang, G. BEVDet4D: Exploit Temporal Cues in Multi-camera 3D Object Detection. arXiv 2022, arXiv:2203.17054. [Google Scholar]
  12. He, Y.; Chen, W.; Xun, T.; Tan, Y. Real-Time 3D Occupancy Prediction via Geometric-Semantic Disentanglement. arXiv 2024, arXiv:2407.13155. [Google Scholar]
  13. Duan, Y.; Wang, W.; Chen, Z.; Zhu, X.; Lu, L.; Lu, T.; Qiao, Y.; Li, H.; Dai, J.; Wang, W. Vision-RWKV: Efficient and Scalable Visual Perception with RWKV-Like Architectures. arXiv 2024, arXiv:2403.02308. [Google Scholar]
  14. Tian, X.; Jiang, T.; Yun, L.; Wang, Y.; Wang, Y.; Zhao, H. Occ3D: A Large-Scale 3D Occupancy Prediction Benchmark for Autonomous Driving. arXiv 2023, arXiv:2304.14365. [Google Scholar]
  15. Zhang, Y.; Zhu, Z.; Du, D. OccFormer: Dual-path Transformer for Vision-based 3D Semantic Occupancy Prediction. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023. [Google Scholar]
  16. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is All you Need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NeurIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  17. Cheng, B.; Misra, I.; Schwing, A.G.; Kirillov, A.; Girdhar, R. Masked-attention Mask Transformer for Universal Image Segmentation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  18. Wei, Y.; Zhao, L.; Zheng, W.; Zhu, Z.; Zhou, J.; Lu, J. SurroundOcc: Multi-Camera 3D Occupancy Prediction for Autonomous Driving. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023. [Google Scholar]
  19. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable detr: Deformable transformers for end-to-end object detection. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual Event, 26–30 April 2020. [Google Scholar]
  20. Jiang, H.; Cheng, T.; Gao, N.; Zhang, H.; Liu, W.; Wang, X. Symphonize 3D Semantic Scene Completion with Contextual Instance Queries. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  21. Huang, Y.; Zheng, W.; Zhang, Y.; Zhou, J.; Lu, J. Tri-Perspective View for Vision-Based 3D Semantic Occupancy Prediction. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023. [Google Scholar]
  22. Lu, Y.; Zhu, X.; Wang, T.; Ma, Y. OctreeOcc: Efficient and Multi-Granularity Occupancy Prediction Using Octree Queries. arXiv 2023, arXiv:2312.03774. [Google Scholar]
  23. Meagher, D. Geometric modeling using octree encoding. Comput. Graph. Image Process. 1982, 19, 129–147. [Google Scholar] [CrossRef]
  24. Ding, X.; Zhang, Y.; Ge, Y.; Zhao, S.; Song, L.; Yue, X.; Shan, Y. Unireplknet: A universal perception large-kernel convnet for audio, video, point cloud, time-series and image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  25. Li, B.; Deng, J.; Zhang, W.; Liang, Z.; Du, D.; Jin, X.; Zeng, W. Hierarchical Temporal Context Learning for Camera-Based Semantic Scene Completion. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 1–4 October 2024. [Google Scholar]
  26. Park, J.; Xu, C.; Yang, S.; Keutzer, K.; Kitani, K.M.; Tomizuka, M.; Zhan, W. Time will tell: New outlooks and a baseline for temporal multi view 3d object detection. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual Event, 1–5 May 2023. [Google Scholar]
  27. Philion, J.; Fidler, S. Lift, splat, shoot: Encoding images from arbitrary camera rigs by implicitly unprojecting to 3d. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020. [Google Scholar]
  28. Li, Y.; Ge, Z.; Yu, G.; Yang, J.; Wang, Z.; Shi, Y.; Sun, J.; Li, Z. Bevdepth: Acquisition of reliable depth for multi-view 3d object detection. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Washington, DC, USA, 7–14 February 2023. [Google Scholar]
  29. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  30. Liu, Y.; Tian, Y.; Zhao, Y.; Yu, H.; Xie, L.; Wang, Y.; Ye, Q.; Liu, Y. VMamba: Visual State Space Model. arXiv 2024, arXiv:2401.10166. [Google Scholar]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  32. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuscenes: A multi modal dataset for autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  33. Loshchilov, I.; Hutter, F. Fixing Weight Decay Regularization in Adam. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  34. Li, Z.; Wang, W.; Li, H.; Xie, E.; Sima, C.; Lu, T.; Qiao, Y.; Dai, J. Bevformer: Learning bird’s-eye-view representation from multi-camera images via spatiotemporal transformers. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  35. Gan, W.; Mo, N.; Xu, H.; Yokoya, N. A Comprehensive Framework for 3D Occupancy Estimation in Autonomous Driving. IEEE Trans. Intell. Veh. 2024. early access. [Google Scholar] [CrossRef]
Figure 1. Comparisons between occupancy prediction temporal fusion methods. (a) illustrates a widely used architecture for temporal fusion, where the backbone extracts all temporal features and sends them to the fusion module for aggregation. (b) is the architecture that GSD-Occ [12] employs. It stores the extracted history features in a queue and retrieves these features from the queue before the temporal fusion process. (c) represents our proposed method. We condense historical features through the Scene Memory Gate and pass the fused historical feature map to the next iteration.
Figure 1. Comparisons between occupancy prediction temporal fusion methods. (a) illustrates a widely used architecture for temporal fusion, where the backbone extracts all temporal features and sends them to the fusion module for aggregation. (b) is the architecture that GSD-Occ [12] employs. It stores the extracted history features in a queue and retrieves these features from the queue before the temporal fusion process. (c) represents our proposed method. We condense historical features through the Scene Memory Gate and pass the fused historical feature map to the next iteration.
Bdcc 09 00176 g001
Figure 2. The overall architecture of ReOcc. The backbone extracts image features from the input, which are then transformed into 3D voxel features via a 2D-to-3D transformation. On the semantic branch, the 3D voxel feature is compressed with a BEV feature. The Scene Memory Gate fuses the BEV feature with the historical feature map, and the fused feature is then fed into the 2D semantic encoder. On the geometric branch, the 3D voxel feature is processed through the 3D geometric encoder. The semantic and geometric features are combined to generate the prediction result.
Figure 2. The overall architecture of ReOcc. The backbone extracts image features from the input, which are then transformed into 3D voxel features via a 2D-to-3D transformation. On the semantic branch, the 3D voxel feature is compressed with a BEV feature. The Scene Memory Gate fuses the BEV feature with the historical feature map, and the fused feature is then fed into the 2D semantic encoder. On the geometric branch, the 3D voxel feature is processed through the 3D geometric encoder. The semantic and geometric features are combined to generate the prediction result.
Bdcc 09 00176 g002
Figure 3. The overview of Scene Memory Gate. The left part shows the iteration process of SMG, and the right part illustrates its detailed architecture. There’s a normalization layer between the convolution layer and the ReLU activation layer, which is omitted for clarity.
Figure 3. The overview of Scene Memory Gate. The left part shows the iteration process of SMG, and the right part illustrates its detailed architecture. There’s a normalization layer between the convolution layer and the ReLU activation layer, which is omitted for clarity.
Bdcc 09 00176 g003
Figure 4. The overview of the Light Encoder. The left part shows the Light Encoder’s architecture, while the right part illustrates the VRWKV architecture.
Figure 4. The overview of the Light Encoder. The left part shows the Light Encoder’s architecture, while the right part illustrates the VRWKV architecture.
Bdcc 09 00176 g004
Figure 5. Qualitative results comparison on Occ3D-NuScenes validation set. Our proposed ReOcc is able to capture more detailed information compared with GSD-Occ. Moreover, ReOcc’s reconstruction is more stable and avoids creating nonexistent structures in the scene (row 1).
Figure 5. Qualitative results comparison on Occ3D-NuScenes validation set. Our proposed ReOcc is able to capture more detailed information compared with GSD-Occ. Moreover, ReOcc’s reconstruction is more stable and avoids creating nonexistent structures in the scene (row 1).
Bdcc 09 00176 g005
Table 1. Three-dimensional occupancy prediction performance on the Occ3D-NuScenes. ’-’ indicates the metrics are not reported. Best values are bolded.
Table 1. Three-dimensional occupancy prediction performance on the Occ3D-NuScenes. ’-’ indicates the metrics are not reported. Best values are bolded.
Method Backbone InputSize Vis . Mask mIoU RayIoU RayIoU 1 m RayIoU 2 m RayIoU 4 m FPS
BEVFormer [34]R1011600 × 90039.232.426.132.938.03.0
FB-Occ(16f) [10]R50704 × 25639.133.526.734.139.710.3
FlashOcc [6]R50704 × 25632.0----29.6
PanoOcc [4]R1011600 × 90042.1----3.0
COTR [3]R50704 × 25644.5----0.9
GSD-Occ(16f) [12]R50704 × 25639.4----20.0
ReOcc-B(2f) (ours)R50704 × 25639.931.525.032.037.623.4
ReOcc-S(2f) (ours)R50704 × 25639.231.424.931.937.326.4
BEVFormer [34]R1011600 × 90023.733.7---3.0
FB-Occ(16f) [10]R50704 × 25627.935.6---10.3
SimpleOcc [35]R101672 × 33631.822.517.022.727.99.7
BEVDet-Occ(2f) [11]R50704 × 25636.129.623.630.035.12.6
BEVDet-Occ-Long(8f) [11]R50 704 × 384 39.332.626.633.138.20.8
SparseOcc(8f) [5]R50704 × 25630.134.028.034.739.417.3
GSD-Occ(16f) [12]R50704 × 25631.838.932.939.744.020.0
Panoptic-FlashOcc(8f) [9]R50704 × 25631.638.532.839.343.435.6
ReOcc-B(2f) (ours)R50704 × 25631.938.232.438.943.223.4
ReOcc-S(2f) (ours)R50704 × 25631.637.831.938.643.026.4
Table 2. The detailed comparison of 3D occupancy prediction performance with real-time methods on the Occ3D-NuScenes. ’-’ indicates the metrics are not reported. ’*’ indicates the metrics are tested by us. Best values are bolded.
Table 2. The detailed comparison of 3D occupancy prediction performance with real-time methods on the Occ3D-NuScenes. ’-’ indicates the metrics are not reported. ’*’ indicates the metrics are tested by us. Best values are bolded.
Method Backbone InputSize Vis . Mask mIoU RayIoU FPS Storage *↓ Param . *↓
GSD-Occ(2f) [12]R50704 × 25637.2-21.02.5 M114.7 M
GSD-Occ(8f) [12]R50704 × 25638.4-20.615.7 M114.8 M
GSD-Occ(16f) [12]R50704 × 25639.4-20.032.3 M115.0 M
ReOcc-B(2f) (ours)R50704 × 25639.931.523.43.5 M59.1 M
ReOcc-S(2f) (ours)R50704 × 25639.231.426.41.9 M55.8 M
SparseOcc(8f) [5]R50704 × 25630.134.017.380.3 M
GSD-Occ(16f) [12]R50704 × 25631.838.920.032.3 M115.0 M
Panoptic-FlashOcc(2f) [9]R50704 × 25630.336.835.96.8 M75.2 M
Panoptic-FlashOcc(8f) [9]R50704 × 25631.638.535.655.0 M76.8 M
Panoptic-FlashOcc(2f) * [9]R50704 × 25630.336.822.5 *75.2 M
Panoptic-FlashOcc(8f) * [9]R50704 × 25631.638.56.9 *76.8 M
ReOcc-B(2f) (ours)R50704 × 25631.938.223.43.5 M59.1 M
ReOcc-S(2f) (ours)R50704 × 25631.637.826.41.9 M55.8 M
Table 3. Ablation on different components. Best values are bolded.
Table 3. Ablation on different components. Best values are bolded.
Method IoU mIoU RayIoU FPS Param .
Baseline67.0736.8529.7123.670.84 M
+SMG69.4339.4331.5223.671.07 M
+Light Encoder68.1337.0229.4423.358.91 M
Final70.5039.9731.5923.459.15 M
Table 4. Ablation on Light Encoder. Best values are bolded.
Table 4. Ablation on Light Encoder. Best values are bolded.
Method Blocks IoU mIoU RayIoU FPS Param .
ResNet270.0739.6931.8924.259.55 M
VMamba [30]270.4439.8931.8622.758.95 M
VRWKV [13]270.5039.9731.5923.459.15 M
VRWKV [13]170.2939.6731.5624.256.81 M
VRWKV [13]370.8740.2031.7122.064.92 M
Table 5. Ablation on temporal fusion methods. Best values are bolded.
Table 5. Ablation on temporal fusion methods. Best values are bolded.
Method IoU mIoU RayIoU FPS Param . Storage
SMG70.5039.9731.5923.459.15 M3.53 M
GSD fusion(2f)68.1337.0229.4423.358.91 M2.53 M
GSD fusion(16f)69.8739.0831.1823.059.06 M25.09 M
Table 6. Ablation on SMG. Best values are bolded.
Table 6. Ablation on SMG. Best values are bolded.
Method IoU mIoU RayIoU RayIoU 1 m RayIoU 2 m RayIoU 4 m
SMG70.5039.9731.5925.0432.0537.66
SMG w/o mask70.2739.5531.1424.7031.6337.08
SMG w/ FG70.5939.7731.4624.8831.9737.54
Table 7. Ablation on 3D-to-BEV transformation. Best values are bolded.
Table 7. Ablation on 3D-to-BEV transformation. Best values are bolded.
Method IoU mIoU RayIoU RayIoU 1 m RayIoU 2 m RayIoU 4 m
mean pooling70.5039.9731.5925.0432.0537.66
convolution70.5139.6431.3824.9331.8137.39
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Z.; Xie, Y.; Wei, Y. RecurrentOcc: An Efficient Real-Time Occupancy Prediction Model with Memory Mechanism. Big Data Cogn. Comput. 2025, 9, 176. https://doi.org/10.3390/bdcc9070176

AMA Style

Chen Z, Xie Y, Wei Y. RecurrentOcc: An Efficient Real-Time Occupancy Prediction Model with Memory Mechanism. Big Data and Cognitive Computing. 2025; 9(7):176. https://doi.org/10.3390/bdcc9070176

Chicago/Turabian Style

Chen, Zimo, Yuxiang Xie, and Yingmei Wei. 2025. "RecurrentOcc: An Efficient Real-Time Occupancy Prediction Model with Memory Mechanism" Big Data and Cognitive Computing 9, no. 7: 176. https://doi.org/10.3390/bdcc9070176

APA Style

Chen, Z., Xie, Y., & Wei, Y. (2025). RecurrentOcc: An Efficient Real-Time Occupancy Prediction Model with Memory Mechanism. Big Data and Cognitive Computing, 9(7), 176. https://doi.org/10.3390/bdcc9070176

Article Metrics

Back to TopTop