Next Article in Journal
Human Movement Recognition Based on 3D Point Cloud Spatiotemporal Information from Millimeter-Wave Radar
Next Article in Special Issue
Meshless Search SR-STAP for Airborne Radar Based on Meta-Heuristic Algorithms
Previous Article in Journal
Correction of Spacecraft Magnetic Field Noise: Initial Korean Pathfinder Lunar Orbiter MAGnetometer Observation in Solar Wind
Previous Article in Special Issue
High-Resolution L-Band TomoSAR Imaging on Forest Canopies with UAV Swarm to Detect Dielectric Constant Anomaly
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SMIFormer: Learning Spatial Feature Representation for 3D Object Detection from 4D Imaging Radar via Multi-View Interactive Transformers

1
School of Automotive Studies, Tongji University, Shanghai 201804, China
2
School of Information Science and Engineering, East China University of Science and Technology, Shanghai 200237, China
3
Shanghai Motor Vehicle Inspection Certification & Tech Innovation Center Co., Ltd., Shanghai 201805, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(23), 9429; https://doi.org/10.3390/s23239429
Submission received: 23 October 2023 / Revised: 5 November 2023 / Accepted: 7 November 2023 / Published: 27 November 2023

Abstract

:
4D millimeter wave (mmWave) imaging radar is a new type of vehicle sensor technology that is critical to autonomous driving systems due to its lower cost and robustness in complex weather. However, the sparseness and noise of point clouds are still the main problems restricting the practical application of 4D imaging radar. In this paper, we introduce SMIFormer, a multi-view feature fusion network framework based on 4D radar single-modal input. SMIFormer decouples the 3D point cloud scene into 3 independent but interrelated perspectives, including bird’s-eye view (BEV), front view (FV), and side view (SV), thereby better modeling the entire 3D scene and overcoming the shortcomings of insufficient feature representation capabilities under single-view built from extremely sparse point clouds. For multi-view features, we proposed multi-view feature interaction (MVI) to exploit the inner relationship between different views by integrating features from intra-view interaction and cross-view interaction. We evaluated the proposed SMIFormer on the View-of-Delft (VoD) dataset. The mAP of our method reached 48.77 and 71.13 in the fully annotated area and the driving corridor area, respectively. This shows that 4D radar has great development potential in the field of 3D object detection.

1. Introduction

The conventional automotive radar has been widely used in advanced driver assistance systems (ADAS) and autonomous driving [1], but it faces limitations such as the absence of elevation information and low resolution compared with LiDAR-based perception technologies [2,3,4,5,6]. To overcome these limitations, the development of 4D imaging radar [7] has emerged as a promising solution. Unlike conventional radar, 4D imaging radar can measure the pitch angle, which allows for the acquisition of elevation information. This additional dimension enhances the understanding of the environment and improves the accuracy of object detection and localization. The literature [8,9] extensively explores the evolution and performance of 4D radar technology, with various studies highlighting advancements and discussing different brands in this field. Short-range sensing applications specifically demonstrate the exceptional performance of 4D radar [10].
When compared with LiDAR, perception systems based on 4D imaging radar offer several advantages. These systems have lower hardware and maintenance costs for object detection, making them more cost-effective. Moreover, they exhibit resilience to external factors. LiDAR point cloud and 4D radar point cloud share similarities, but they also have notable differences; 4D imaging radar enables velocity measurements, enhancing the original point cloud data by providing additional dimensions to represent the vector velocity information of each point. While both technologies provide valuable data for perception systems, LiDAR’s inability to penetrate fog and dust hinders its application in severe weather [11], and 4D radar measurements are subject to noise due to multi-path propagation and limited angular resolution caused by longer wavelengths and a restricted number of antennas. This limited number of antennas in 4D radar results in a sparser point cloud compared with the denser point cloud produced by LiDAR systems [12]. Consequently, 4D radar captures less geometry and semantic information compared with LiDAR. This difference in point cloud density poses a challenge for existing 3D object detection algorithms, which are specifically designed for dense LiDAR point clouds. As a result, these algorithms may not perform optimally when directly applied to sparse 4D radar point cloud data.
Currently, methods based on LiDAR point clouds often employ voxel-based representations, where the 3D space is divided into voxels and each voxel is assigned a vector to indicate its status. However, the sheer number of voxels presents computational challenges, necessitating specialized techniques such as sparse convolution [13]. In outdoor scenes, the distribution of information is not uniform. To address this, modern methods based on LiDAR point clouds collapse the height dimension and focus on the bird’s-eye view (BEV), as this perspective exhibits the most variability in information [14,15,16,17,18,19,20]. These methods encode the 3D information of each object in the BEV grid. BEV-based methods excel in LiDAR 3D object detection tasks due to their ability to predict coarse-level bounding boxes for commonly encountered objects [16,17]. Furthermore, as depicted in Figure 1, the dense nature of the Lidar point cloud allows for a more comprehensive acquisition of an object’s surface profile compared with the 4D radar point cloud. On average, objects in a 4D imaging radar point cloud can only be assigned to less than 200 points, which poses a challenge in encoding the extremely sparse outline information of objects with various 3D structures using a flattened vector.
To develop a safer and more robust 4D radar autonomous driving system, a more comprehensive and fine-grained understanding of the 3D surroundings is necessary. The generalization of BEV to model fine-grained 3D structures while maintaining efficiency and detection performance remains an open question. In other words, the use of spatial geometry information has become a key factor in improving the accuracy and performance of the 3D object detection field of 4D radar. In order to enhance 3D surrounding perception, it is essential to effectively represent 3D scenes.
In this paper, we propose a method that aims to retain and incorporate multi-view geometric information into spatial features for more robust object detection. Recognizing the untapped potential within multi-view geometric data, we seek to bridge this gap by incorporating it into spatial features for object detection purposes. By considering multiple views simultaneously, we aim to capture a more comprehensive understanding of objects from different angles or orientations. This approach holds great promise for improving both accuracy and performance across diverse scenarios. We take inspiration from previous research but introduce novel techniques specifically designed to retain and incorporate multi-view geometric information effectively. By integrating data from multiple viewpoints during feature extraction processes, we create richer representations that enable more precise identification and localization of objects within complex scenes.
By leveraging multi-view geometric information, our method offers several potential benefits towards achieving enhanced accuracy or improved overall performance in object detection tasks: Robustness: Considering multiple views helps mitigate occlusion issues commonly encountered when analyzing 3D scenes. By capturing information from various angles, our method can better handle instances where objects may be partially or fully obscured from certain perspectives. Discriminative Power: The inclusion of multi-view geometric features provides additional discriminative power in distinguishing between objects that may share similar appearances but differ in their spatial characteristics. This added information aids in reducing false positives and improving detection precision. Generalization: Incorporating multi-view perspectives promotes greater generalization capabilities, enabling the model to adapt more effectively to unseen or novel scenarios by leveraging a broader range of spatial representations. The main contributions of our work can be summarized as follows:
  • We addressed the issue of insufficient representation of extremely sparse point clouds in a single view by decoupling 3D voxel features into separate front view (FV), side view (SV), and bird’s-eye view (BEV) planes.
  • We propose sparse-dimensional compression as an alternative to dense-dimensional compression. By individually placing voxels on each view plane and aggregating features at corresponding positions, we construct a two-dimensional sparse feature matrix. This approach allows us to achieve precise predictions while minimizing memory and computational demands.
  • We suggest multi-view feature interaction (MVI) as a method to enhance spatial perception. MVI divides the full-size feature map into non-overlapping windows, enabling more effective interaction between view-inside and view-outside features through self-attention and cross-attention. This approach enhances spatial perception at each interaction level with only a small increase in computation.

2. Related Work

2.1. 3D Object Detection with LiDAR Point Cloud

Voxel-based and pillar-based feature representations are the main methods of relevant prior grid-based 3D detection work using LiDAR point clouds.
VoxelNet [21] stands as one of the pioneering studies that employ end-to-end 3D detection by dividing a point cloud into voxels and accurately predicting 3D bounding boxes. While it offers promising results in object recognition tasks, its high computational cost limits its applicability for real-time applications. SECOND [22] introduces an innovative approach through the utilization of 3D sparse convolution to enhance performance while reducing computational overheads. By extracting voxel features using a backbone network with sparse convolution techniques, these features are then concatenated along the height dimension before undergoing additional processing via 2D convolution layers to obtain dense bird’s-eye view (BEV) features. This methodology significantly accelerates computation speed without compromising accuracy. Building upon the foundations laid by SECOND [22], CenterPoint [2] proposes a novel idea of employing a single positive cell for each detected object. By refining existing methodologies further, CenterPoint aims to streamline detection pipelines by optimizing computations involved in identifying individual objects within crowded scenes effectively. The method known as VISTA [23] takes another step forward by projecting 3D feature maps onto both bird’s-eye view (BEV) and range view (RV). Through this multi-view transformer fusion approach, VISTA achieves enhanced mapping capabilities for improved object recognition across different perspectives. Recognizing the importance of foreground data during learning processes led researchers to develop FocalsConv, a dynamic mechanism that dynamically focuses on foreground information [24]. By selectively emphasizing relevant data, FocalsConv aims to refine the learning process and enhance detection accuracy. Transformer [25] is a powerful global modeling method. CenterFormer [26] improves the prediction accuracy of bounding boxes by aggregating features around the center candidate and introducing this architecture into the detection head. The majority of existing methods (specifically [2,23,24,27,28]) have been developed based on the framework established by SECOND [22]. While these methods have shown promise in various object detection tasks, they fall short in effectively leveraging multi-view geometric information. By solely relying on single perspectives or views, these approaches overlook potentially significant details that could enhance their accuracy and overall performance.
PointPillars [29] is a pioneering pillar-based approach that utilizes PointNets [30] to encode point features before transforming them into a pseudo-image in bird’s-eye view (BEV) using pooling operations. By employing only 2D convolutional layers, it enables end-to-end learning while making it suitable for computation-constrained embedded systems with low latency requirements. Infofocus [31] enhances PointPillars [29] by introducing a second-stage attention network specifically designed for fine-grained proposal refinement. This additional network improves precision and accuracy by focusing on important details within detected objects. Based on CenterPoint-pillar [2] architecture, PillarNet [32] incorporates the ResNet18 structure with 2D sparse convolution into its backbone for BEV feature extraction. This modification further enhances computational efficiency while maintaining high performance levels. Experiments have demonstrated that after sufficient extraction through 2D sparse convolution techniques, the accuracy achieved by pillar-based networks can be comparable to voxel-based methods commonly used in 3D object detection tasks. However, this accomplishment comes at a cost; pillar-based approaches suffer from significant loss of critical 3D geometric information during transformation processes. Due to the inherent nature of pillarization and subsequent projection onto BEV representations, these approaches struggle to overcome limitations associated with preserving essential aspects related to 3D object detection. The loss of geometric information poses challenges in accurately localizing objects and understanding their spatial relationships within the scene.

2.2. 3D Object Detection with 4D Imaging Radar Point Cloud

In the past period of time, despite the fact that the traditional radar was cheaper than lidar and was robust in complex weather, it was unable to provide sufficient support for 3D object detection due to its point cloud being too sparse and lacking elevation information. However, with the maturity of 4D millimeter wave imaging radar technology, the density of radar point clouds has been significantly improved, and it can provide elevation information like lidar and additional speed information. This allows people to see the potential of using 4D radar for 3D object detection.
Based on pillarization inspired by PointPillars [29], RPFA-Net [33] designs a self-attention module for sparse columnar features to better globally model sparse features. RadarMFNet [34] compensates the speed information of the radar point cloud by estimating the speed information of the ego vehicle and matches the multi-frame millimeter wave imaging radar point cloud to the last frame, using a multi-frame radar point cloud for 3D object detection. SMURF [35] Introducing kernel density estimation into radar point clouds, using additional statistical features to effectively mitigate the adverse effects of inherent noise and sparsity in point clouds. Furthermore, research on multi-modal fusion of camera images has also been applied to radar point clouds. RCBEV [36] uses the camera image as the forward view (FV) and the radar point cloud as the BEV for multi-view fusion. CRN [37] uses radar point clouds for space occupancy prediction to better guide the conversion of camera image features into BEV features. RCFusion [38] lifts camera image features to 3D radar point cloud coordinates through orthogonal feature transformation (OFT) to achieve multi-modal fusion. LXL [39] uses sampling-based methods to lift 2D image features to 3D point cloud space and further optimize the lifted features through depth prediction and occupancy prediction.

2.3. Point Cloud Perception for Indoor Scenes

The change in feature extraction method brought about the change in point cloud density can also be extended to indoor scenes. Indoor scenes tend to have a higher density of point clouds than outdoor lidar, let alone 4D imaging radar. That is, we can gain insight into how the 3D scene is represented by observing the difference in the way features are extracted as point cloud sparsity decreases.
Poly-PC [40] proposes a Res-SA layer for efficient scaling and a weight-entanglement-based one-shot NAS technique to find optimal architectures, with a task-prioritization-based gradient balance algorithm for training and high performance in various tasks. CO-Net [41] uses a Res-MLP block for feature extraction, a nested layer-wise processing policy for optimal architecture, and sign-based gradient surgery to promote training and optimize task-shared parameters in multi-task learning on point clouds. FARP-Net [42] introduces an LGFAL layer that combines local and global features using an attention-based fusion module, incorporates an LFAM to map local features into a normal distribution, and proposes a WRPM for object proposal generation that weighs relation importance among object candidates for improved proposal quality. Auto-Points [43] uses a SAM layer for flexible scaling, simplifies architecture search with child candidates, and optimizes them using a weight-entwinement NAS technique for each point cloud task.

3. Method

In LiDAR-centric autonomous driving perception tasks, it is common to use a 3D convolutional backbone network for extracting features from the point cloud. These features are then inputted into a dedicated encoder based on the presentation framework. To capture more spatial context information in sparser 4D imaging radar point clouds (over 200 times sparser than LiDAR point clouds), we introduce SMIFormer. Figure 2 showcases the structure of SMIFormer, which proposes the utilization of a transformer-based multi-view encoder. This encoder utilizes the attention mechanism to enhance the point cloud feature in the multi-view plane.

3.1. Framework Overview

In SMIFormer, we propose the incorporation of decoupled per-view feature encoding, feature self-attention inside the view, and feature cross-attention outside the view. These techniques are employed to efficiently extract multi-view plane features. Each view plane represents a mesh cell feature associated with one of the three planes, capturing specific view information from the corresponding pillar area. The feature self-attention inside the view focuses on encoding in-view features by interacting with view planar features within the same view. On the other hand, feature cross-attention outside the view facilitates direct interaction between view plane features across different views, allowing for the incorporation of richer contextual information from all perspectives.

3.2. Decoupled Per-View Feature Encoding

To generate a comprehensive observation of the scene from different perspectives and overcome limitations such as object occlusion caused by solely extracting bird’s-eye view features, we compress the voxel features into separate front view (FV), side view (SV), and bird’s-eye view (BEV) planes when constructing the three-view features.
The 3D object detectors presented in papers [2,22,29] employ compression techniques to convert sparse 3D voxel features into dense 2D maps. This is achieved by transforming sparse features into dense ones and incorporating altitude information (along the z-axis) into the channel dimension. However, these operations require additional memory and computational resources.
In order to achieve precise predictions while minimizing memory and computational demands, we utilize sparse dimension compression to separate the features of each view. This involves placing voxels individually on each view plane and aggregating the features at corresponding positions to construct a two-dimensional sparse feature matrix ( D b e v , D f v , D s v ) . The construction of this matrix relies on the values of sparse features and their coordinate indexes. After performing sparse dimension compression, the calculation of sparse features F v and their corresponding coordinates ( C b e v , C f v , C s v ) is as follows:
C b e v = x c , y c , 0 c C v o x e l C f v = x c , 0 , z c c C v o x e l C s v = 0 , y c , z c c C v o x e l
F v = c A c ^ f c , c ^ C v , v b e v , f v , s v A c ^ = c x c = x c ^ , y c = y c ^ , z c = z c ^ , c C v

3.3. Voxel Feature Query Module

The decoupled multi-view representation aims to provide a comprehensive description of a query voxel positioned at ( x , y , z ) in three-dimensional space. This is accomplished by consolidating its projections on BEV, SV, and FV views, as illustrated in Figure 3.
Specifically, we initiate the process by projecting the voxels onto the BEV, SV, and FV view planes, thereby obtaining the coordinates [ ( x , y , 0 ) , ( x , 0 , z ) , ( 0 , y , z ) ] . Subsequently, we sample the BEV, SV, and FV planes at these coordinates to retrieve the corresponding features [ f b e v ( x , y , 0 ) , f s v ( x , 0 , z ) , f f v ( 0 , y , z ) ] :
f b e v ( x , y , 0 ) = I ( F b e v , T b e v ( x , y , 0 ) ) f s v ( x , 0 , z ) = I ( F s v , T s v ( x , 0 , z ) ) f f v ( 0 , y , z ) = I ( F f v , T f v ( 0 , y , z ) )
By combining these three features with the voxel’s intrinsic features, we generate the final f s p a t i a l ( x , y , z ) :
f s p a t i a l ( x , y , z ) = S U M ( f v o x e l ( x , y , z ) , f b e v ( x , y , 0 ) , f s v ( x , 0 , z ) , f f v ( 0 , y , z ) )
The variables F b e v , F s v , and F f v are feature maps represented by tensors with dimensions R W b e v × H b e v × D , R W f v s v × H f v s v × D , and R W f v s v × H f v s v × D , respectively. The variable f is a feature vector represented by a tensor with dimensions R 1 × D . I represent the index function that samples the feature vectors from the feature maps of BEV, SV, and FV based on the given coordinates. The coordinate transformation function, denoted as T, is defined as follows: Considering that the BEV, SV, and FV view planes align with the axes of three-dimensional space, each projection function T only needs to perform a straightforward index dimensionality reduction on the two associated coordinate systems it encompasses.

3.4. Splitting for Multi-View Feature Interaction

The aforementioned decoupled per-view feature encoding structure compresses the sparse three-dimensional feature voxels into three views, utilizing the feature map as the representation of each perspective to enable multi-perspective feature interaction. However, due to the high-resolution nature of each view plane feature (approximately 10 5 queries), it becomes inefficient and feature-redundant to compute full-scale attention directly within the same view, considering the significant computational costs and GPU memory requirements. To address this issue, we propose a windowing strategy to temporally divide the full-size feature map into non-overlapping windows, with each window covering a small area of the feature map. Specifically, the two-dimensional sparse feature map with spatial dimensions W × H is evenly divided into g = g w × g h groups, where each group has spatial dimensions W × H (assuming W = W × g w , H = H × g h ). This paper’s windowing strategy enables more effective multi-view feature interaction through the proposed self-attention of view-inside features and cross-attention of view-outside features.

3.5. Feature Self-Attention inside the View

In SMIFormer, we utilize the self-attention of view-inside features to promote feature interaction within each view. To achieve this, we first gather sparse features F i = { f 1 , i , , f N i , i } within each window i for a given perspective’s feature map. Each window can be represented as a three-dimensional sparse feature matrix. The spatial shape of these matrices varies depending on the type of view (bird’s-eye view or front/side views). For example, in a bird’s-eye view representation, the spatial shape is W b e v × H b e v × D , while it is W f v s v × H f v s v × D for front and side views. Here, D represents the feature dimensionality per non-empty voxel. In the case of the bird’s-eye view, the features F i within each window form an N i × D matrix in its three-dimensional sparse matrix representation. It is important to note that these three-dimensional feature matrices are sparse, thus satisfying the condition ( W i × H i ) N i . To further facilitate interaction among all sparse voxels within each view, we calculate feature multi-head self-attention within the view on a per-view basis:
F ^ i = M A L q F i + P E , L k F i + P E , L v F i
The linear projection layers, L q ( · ) , L k ( · ) , and L v ( · ) , are employed to generate query, key, and value features for the multi-head self-attention mechanism while maintaining the same semantics. To integrate spatial information into the attention process, we introduce ‘PE’ as a learnable index-based 2D position embedding for each sparse voxel. This embedding is obtained by mapping the indices ( i , j ) of the sparse voxels through a Multi-Layer Perceptron (MLP).

3.6. Feature Cross-Attention outside the View

We aim to propagate sparse voxel information at the view-level across different views after applying self-attention within the view. This allows us to capture more comprehensive three-dimensional spatial information. To achieve this, we employ the attention mechanism using the per-view feature representation. We query the features summarized in the full view of the entire three-dimensional space.
To enhance the fusion of features between each view, we calculate the cross-attention of the external features in the view. Specifically, we group all the divided windows in all views based on their voxel projection in 3D space. When windows from the same voxel block are projected onto the three views, they are grouped together. This grouping enables the interaction of window features from the three views within the group, further enriching the features of each view with information from other views. In the following example, a view is used as the query feature, while other views serve as the key and value features for calculating multi-head cross-attention:
F ^ i v 1 = M A L q F i v 1 + P E v 1 , L k F i v 2 + P E v 2 , L v F i v 2 + M A L q F i v 1 + P E v 1 , L k F i v 2 + P E v 3 , L v F i v 3
Each view performs similar calculations to update its own features.

4. Experiments

4.1. Dataset

In this study, we utilize the View of Delft (VoD) [44] dataset to evaluate the performance of the proposed method. This dataset is designed to validate the performance of radar-based road detection. The dataset also covers multi-modal data from various sensors, including 4D radar points, LiDAR points, and camera images. The VoD dataset provides a corresponding label for each object to be detected, including its category, a 3D bounding box, and a tracking ID. As well as coordinate transformation matrices between different sensors.
In our experiment, we mainly detected three main categories in the dataset, including cars, pedestrians, and cyclists. For the division of the data set, we follow the official split method. Specifically, the VoD dataset comprises 5139 frames for training and 1296 frames for validation. Since the VoD data set does not provide test set labels and automated test set evaluation systems, our tests and analysis are performed on the validation set.

4.2. Evaluation Metrics

We evaluated SMIFormer using mean average precision (mAP). The threshold used in the calculation of AP is based on the intersection over union (IoU) of the predicted and ground truth bounding boxes in 3D, which requires an overlap of 0.5 , 0.25 , and 0.25 for cars (Car), pedestrians (Ped), and cyclists (Cyc), respectively. In the VoD dataset, there are two official evaluation matrices, each specifying different ranges of detection areas: (1) region of the entire annotation area (camera FoV up to 50 m). (2) region of the driving corridor, defined as a rectangle on the ground plane in front of the vehicle as [−4 m < x < 4 m, z < 25 m] in camera coordinates, which is more relevant to actual driving.

4.3. Implementation Details

The model implementation is based on OpenPCDet [45], an open-source framework designed for 3D object detection, and was trained on four NVIDIA RTX A6000 GPUs for 80 epochs; the batch size is set to 4. The point cloud range (PCR) is limited to D ( P C R ) = { ( x , y , z ) 0 < x < 51.2 m, −25.6 m < y < 25.6 m, −3 m < z < 2 m} in the radar coordinate. The voxel size in the voxelization process for the radar cloud points is set to 0.05 m × 0.05 m × 0.125 m. We implemented Adam as our optimizer, set the initial learning rate to 0.003, weight decay to 0.01 , and momentum to 0.9 . After we use the proposed method to encode the point cloud data and obtain the BEV feature map, We implement the dense head based on the CenterPoint [2] to output the detected bounding boxes. For data augmentation, because the velocity measured by the radar must remain relative to the angle of the observed object, rotating the bounding boxes and the points inside will change the radial component of the object’s velocity in an unknown way [44]. Therefore, we only use mirror inversion and scaling as augmentation.

4.4. Results and Analysis

In this study, we compare SMIFormer with other state-of-the-art models for 3D detection based on point clouds. The experimental results of the VoD [44] validation set are presented in Table 1. To obtain a comprehensive comparison, We deployed some 3D object detection networks designed for LiDAR point clouds on the VoD dataset for evaluation, including anchor-based detectors like PointPillars [29], SECOND [22], and the anchor-free detector CenterPoint [2]. Furthermore, we compare the proposed method with recently proposed 3D object detection methods designed for 4D radar, such as LXL-R [39]. Recently, a multi-modal fusion of camera images and radar point clouds has become mainstream for 3D object detection; thus, we also compared our method with the latest RCFusion [38], which was designed for the feature fusion of 4D radar point clouds and camera images.
Our SMIFormer achieves better results compared with several other detection methods. Compared with CenterPoint, which also uses anchor-free detectors, the proposed method with a three-view fusion encoder has achieved significant improvements, with an increase of 3.35% and 6.07% on mAP across the entire annotation area and driving corridor area. Compared with conventional methods using anchor-based detectors, the proposed method also achieved an improvement of no less than 3.36% and 4.41% in mAP in the two annotated areas, respectively. LXL is a relatively new 3D target detection method based on a 4D radar point cloud, which has achieved optimal results in single-modal detection; our method still outperforms it in terms of mAP, which confirms that our proposed method indeed improves object detection accuracy. At the same time, we found that SMIFormer’s pedestrian detection accuracy is significantly higher than other solutions. This may be because the pedestrian’s bounding boxes are prone to interference and occlusion problems due to their small size. SMIFormer improves pedestrian detection by introducing additional view capabilities to enhance point cloud features associated with small objects.
Despite the fact that SMIFormer only uses a 4D radar point cloud as a single-modal input, compared with the newer radar-camera fusion method RCFusion [38], although the mAP has declined in the entire annotation area, it has achieved significant improvements in the driving corridor area. This may be due to the sparsity of radar points, resulting in some targets’ bounding boxes far away from the radar having very few radar points. The fusion-based detection method uses camera images to supplement the features of these areas, thereby improving the mAP of the entire area. While SIMFormer combines features from different views, it enables the model to have better perception capabilities in relatively dense areas of point clouds, thereby improving detection accuracy in the driving corridor areas.
We visualize the results of SMIFormer on the vod validation set in Figure 4. We can find that the proposed method can detect the target object in the vast majority of cases, which shows that under the support of the tri-perspective fusion, our method can accurately locate the position of the object. For a small number of undetected objects, we observed that there are few or even no point clouds in these ground truth bounding boxes, which results in the generation of empty voxels and prevents the model from obtaining useful features.

4.5. Ablation Study

In this subsection, we performed several experiments on the VoD validation dataset to further demonstrate the effectiveness of our SMIFormer.

4.5.1. Effects of Proposed Components

A comprehensive set of ablation studies is conducted for each of our proposed components, as illustrated in Table 2. VoxelNet [21] is adopted as the baseline, and the Inception sparse convolution block and 2D sparse downsample encoder are consistently employed across all experiments.
Based on the results presented in Table 2, it is evident that the introduction of multi-view plane decoupling into the voxel encoder (MVD, #1 and #2) brings noticeable improvements compared with the baseline. This serves as evidence that incorporating semantic features from different perspective views in the voxel space significantly enhances the model’s ability to represent radar point clouds. Furthermore, the utilization of attention operations further enhances the efficiency of multi-view fusion. Feature self-attention inside the view (SAI, #3) can enhance the global representation of a single view, thereby improving the performance of the network. Additionally, feature cross-attention outside the view (CAO, #4) enhances the network’s ability to represent different perspectives by enabling features from different views to interact in the same semantic space. This integration of features from different perspectives leads to a better overall performance. By combining all of these components (#4), we observe a significant improvement in mAP for both the entire annotation area (from 46.48% to 48.77%) and the driving corridor area (from 68.36% to 71.13%). These results serve as strong evidence for the effectiveness of our proposed method.

4.5.2. Effects of Splitting Window Size

We also investigate the impact of splitting window size on the implementation of feature cross-attention outside the view during the fusion of different views. Window sizes of 2, 4, and 8 are selected, and the model containing all components is evaluated accordingly. The results are presented in Table 3, revealing that the model achieves the highest mAP in both annotation areas when the window size is set to 4. When the window size is reduced to 2, the model’s performance experiences a slight decrease, but it should be noted that a smaller window size leads to increased computational resource requirements during attention calculation per unit of time. Conversely, when the window size is increased to 8, the model encounters a significant drop in performance, particularly in pedestrian detection. This can be attributed to the relatively small bounding boxes of pedestrians. Larger window sizes tend to collect more noise information, thereby affecting the accuracy of detecting small targets.
We conducted an experiment to investigate the impact of different window sizes on inference latency. Specifically, we tested the inference time of SMIFormer using various sizes of splitting windows on a single GPU. The results are presented in Table 4. It is evident that both too-small and too-large window sizes result in increased reasoning delay. When the window size is too small, the number of generated windows increases, leading to a higher frequency of attention calculations and thus longer inference times. Conversely, when the window size is too large, although the total number of windows decreases, the number of query vectors within each window significantly increases, resulting in increased inference latency. Therefore, in our other experiments, we consistently adopted a splitting window size of 4.

5. Conclusions

In this study, we proposed SMIFormer, a 3D object detection network based on 4D imaging radar. Compared with the traditional 3D object detection backbone network that converts 3D scene features into bird’s-eye view features, due to the sparsity of radar point clouds, the features of a single perspective will also be very sparse and cannot fully represent the whole 3D scene. SMIFormer additionally decouples the front view and side view features, significantly improves the spatial representation capability of the network, and introduces additional downsampling layers to enable the network to extract more deep semantic information. For the interaction between different views, we propose the MVI module, which extracts richer semantic information from sparse point clouds by integrating self-attention-based intra-view feature interaction and cross-attention-based inter-view feature interaction.
In order to solve the problem of the high computational cost of attention mechanism in high-resolution feature maps, we propose a window calculation strategy that divides the multi-view feature maps into a series of non-overlapping windows. Self-attention is performed on internal sparse features in each window, and cross-attention is performed between multi-view windows only when these windows are projected onto the same voxel. This strategy reduces the amount of calculation while ensuring the effectiveness of feature extraction, thereby improving the accuracy of 3D object detection.
The proposed SMIFormer proposes a new method to enhance feature extraction of extremely sparse point clouds, which may be a potential solution for feature extraction of extremely sparse point clouds. Future works will focus on further improving the network’s feature extraction capabilities from sparse point clouds and utilizing camera images to improve multi-view representation of 4D imaging radar point clouds, which will further improve the accuracy and efficiency of 3D object detection based on 4D radar point clouds.

Author Contributions

Conceptualization, Z.Z., K.Z., Y.Z. and W.S.; methodology, Z.Z. and K.Z.; software, Z.Z. and K.Z.; validation, Z.Z. and K.Z.; formal analysis, Z.Z., K.Z., Y.Z. and W.S.; investigation, W.S. and H.C.; resources, H.C., Y.Z. and Z.Y.; data curation, H.C., Y.Z. and Z.Y.; writing—original draft preparation, Z.Z. and K.Z.; writing—review and editing, Z.Z., K.Z. and Y.Z.; visualization, Z.Z. and K.Z.; supervision, H.C., Y.Z., Z.Y. and W.S.; project administration, H.C., Y.Z. and Z.Y.; funding acquisition, H.C., Y.Z. and Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Shanghai Automotive Industry Science and Technology Development Foundation (2304).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors would like to express their gratitude to the anonymous reviewers and editors who worked selflessly to improve our manuscript.

Conflicts of Interest

Author Huanlei Chen was employed by the company Shanghai Motor Vehicle Inspection Certification & Tech Innovation Center Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Sun, S.; Petropulu, A.P.; Poor, H.V. MIMO radar for advanced driver-assistance systems and autonomous driving: Advantages and challenges. IEEE Signal Process. Mag. 2020, 37, 98–117. [Google Scholar] [CrossRef]
  2. Yin, T.; Zhou, X.; Krahenbuhl, P. Center-based 3d object detection and tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11784–11793. [Google Scholar]
  3. Liu, J.; Bai, L.; Xia, Y.; Huang, T.; Zhu, B.; Han, Q.L. GNN-PMB: A simple but effective online 3D multi-object tracker without bells and whistles. IEEE Trans. Intell. Veh. 2022, 8, 1176–1189. [Google Scholar] [CrossRef]
  4. Dreher, M.; Erçelik, E.; Bänziger, T.; Knoll, A. Radar-based 2D car detection using deep neural networks. In Proceedings of the 2020 IEEE 23rd International Conference on Intelligent Transportation Systems (ITSC), Rhodes, Greece, 20–23 September 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–8. [Google Scholar]
  5. Svenningsson, P.; Fioranelli, F.; Yarovoy, A. Radar-pointgnn: Graph based object recognition for unstructured radar point-cloud data. In Proceedings of the 2021 IEEE Radar Conference (RadarConf21), Atlanta, GA, USA, 8–14 May 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–6. [Google Scholar]
  6. Bai, J.; Li, S.; Huang, L.; Chen, H. Robust detection and tracking method for moving object based on radar and camera data fusion. IEEE Sens. J. 2021, 21, 10761–10774. [Google Scholar] [CrossRef]
  7. Han, Z.; Wang, J.; Xu, Z.; Yang, S.; He, L.; Xu, S.; Wang, J. 4D Millimeter-Wave Radar in Autonomous Driving: A Survey. arXiv 2023, arXiv:2306.04242. [Google Scholar]
  8. Brisken, S.; Ruf, F.; Höhne, F. Recent evolution of automotive imaging radar and its information content. IET Radar Sonar Navig. 2018, 12, 1078–1081. [Google Scholar] [CrossRef]
  9. Li, G.; Sit, Y.L.; Manchala, S.; Kettner, T.; Ossowska, A.; Krupinski, K.; Sturm, C.; Lubbert, U. Novel 4D 79 GHz radar concept for object detection and active safety applications. In Proceedings of the 2019 12th German Microwave Conference (GeMiC), Stuttgart, Germany, 25–27 March 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 87–90. [Google Scholar]
  10. Li, G.; Sit, Y.L.; Manchala, S.; Kettner, T.; Ossowska, A.; Krupinski, K.; Sturm, C.; Goerner, S.; Lübbert, U. Pioneer study on near-range sensing with 4D MIMO-FMCW automotive radars. In Proceedings of the 2019 20th International Radar Symposium (IRS), Ulm, Germany, 26–28 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–10. [Google Scholar]
  11. Bansal, K.; Rungta, K.; Zhu, S.; Bharadia, D. Pointillism: Accurate 3d bounding box estimation with multi-radars. In Proceedings of the 18th Conference on Embedded Networked Sensor Systems, Virtual, 16–19 November 2020; pp. 340–353. [Google Scholar]
  12. Bagloee, S.A.; Tavana, M.; Asadi, M.; Oliver, T. Autonomous vehicles: Challenges, opportunities, and future implications for transportation policies. J. Mod. Transp. 2016, 24, 284–303. [Google Scholar] [CrossRef]
  13. Choy, C.; Gwak, J.; Savarese, S. 4d spatio-temporal convnets: Minkowski convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3075–3084. [Google Scholar]
  14. Huang, J.; Huang, G.; Zhu, Z.; Ye, Y.; Du, D. Bevdet: High-performance multi-camera 3d object detection in bird-eye-view. arXiv 2021, arXiv:2112.11790. [Google Scholar]
  15. Li, Y.; Ge, Z.; Yu, G.; Yang, J.; Wang, Z.; Shi, Y.; Sun, J.; Li, Z. Bevdepth: Acquisition of reliable depth for multi-view 3d object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 1477–1485. [Google Scholar]
  16. Liang, T.; Xie, H.; Yu, K.; Xia, Z.; Lin, Z.; Wang, Y.; Tang, T.; Wang, B.; Tang, Z. Bevfusion: A simple and robust lidar-camera fusion framework. Adv. Neural Inf. Process. Syst. 2022, 35, 10421–10434. [Google Scholar]
  17. Liu, Z.; Tang, H.; Amini, A.; Yang, X.; Mao, H.; Rus, D.L.; Han, S. Bevfusion: Multi-task multi-sensor fusion with unified bird’s-eye view representation. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 2774–2781. [Google Scholar]
  18. Philion, J.; Fidler, S. Lift, splat, shoot: Encoding images from arbitrary camera rigs by implicitly unprojecting to 3d. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XIV 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 194–210. [Google Scholar]
  19. Zhang, Y.; Zheng, W.; Zhu, Z.; Huang, G.; Lu, J.; Zhou, J. A simple baseline for multi-camera 3d object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 3507–3515. [Google Scholar]
  20. Zhang, Y.; Zhu, Z.; Zheng, W.; Huang, J.; Huang, G.; Zhou, J.; Lu, J. Beverse: Unified perception and prediction in birds-eye-view for vision-centric autonomous driving. arXiv 2022, arXiv:2205.09743. [Google Scholar]
  21. Zhou, Y.; Tuzel, O. Voxelnet: End-to-end learning for point cloud based 3d object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4490–4499. [Google Scholar]
  22. Yan, Y.; Mao, Y.; Li, B. Second: Sparsely embedded convolutional detection. Sensors 2018, 18, 3337. [Google Scholar] [CrossRef] [PubMed]
  23. Deng, S.; Liang, Z.; Sun, L.; Jia, K. Vista: Boosting 3d object detection via dual cross-view spatial attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8448–8457. [Google Scholar]
  24. Chen, Y.; Li, Y.; Zhang, X.; Sun, J.; Jia, J. Focal sparse convolutional networks for 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5428–5437. [Google Scholar]
  25. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar] [CrossRef]
  26. Zhou, Z.; Zhao, X.; Wang, Y.; Wang, P.; Foroosh, H. Centerformer: Center-based transformer for 3d object detection. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 496–513. [Google Scholar]
  27. Deng, J.; Shi, S.; Li, P.; Zhou, W.; Zhang, Y.; Li, H. Voxel r-cnn: Towards high performance voxel-based 3d object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 1201–1209. [Google Scholar]
  28. Mao, J.; Xue, Y.; Niu, M.; Bai, H.; Feng, J.; Liang, X.; Xu, H.; Xu, C. Voxel transformer for 3d object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 3164–3173. [Google Scholar]
  29. Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. Pointpillars: Fast encoders for object detection from point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12697–12705. [Google Scholar]
  30. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  31. Wang, J.; Lan, S.; Gao, M.; Davis, L.S. Infofocus: 3d object detection for autonomous driving with dynamic information modeling. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part X 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 405–420. [Google Scholar]
  32. Shi, G.; Li, R.; Ma, C. Pillarnet: Real-time and high-performance pillar-based 3d object detection. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 35–52. [Google Scholar]
  33. Xu, B.; Zhang, X.; Wang, L.; Hu, X.; Li, Z.; Pan, S.; Li, J.; Deng, Y. RPFA-Net: A 4D radar pillar feature attention network for 3D object detection. In Proceedings of the 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), Indianapolis, IN, USA, 19–22 September 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 3061–3066. [Google Scholar]
  34. Tan, B.; Ma, Z.; Zhu, X.; Li, S.; Zheng, L.; Chen, S.; Huang, L.; Bai, J. 3d object detection for multi-frame 4d automotive millimeter-wave radar point cloud. IEEE Sens. J. 2022, 23, 11125–11138. [Google Scholar] [CrossRef]
  35. Liu, J.; Zhao, Q.; Xiong, W.; Huang, T.; Han, Q.L.; Zhu, B. SMURF: Spatial Multi-Representation Fusion for 3D Object Detection with 4D Imaging Radar. arXiv 2023, arXiv:2307.10784. [Google Scholar] [CrossRef]
  36. Zhou, T.; Chen, J.; Shi, Y.; Jiang, K.; Yang, M.; Yang, D. Bridging the view disparity between radar and camera features for multi-modal fusion 3d object detection. IEEE Trans. Intell. Veh. 2023, 8, 1523–1535. [Google Scholar] [CrossRef]
  37. Kim, Y.; Kim, S.; Shin, J.; Choi, J.W.; Kum, D. Crn: Camera radar net for accurate, robust, efficient 3d perception. arXiv 2023, arXiv:2304.00670. [Google Scholar]
  38. Zheng, L.; Li, S.; Tan, B.; Yang, L.; Chen, S.; Huang, L.; Bai, J.; Zhu, X.; Ma, Z. RCFusion: Fusing 4D Radar and Camera with Bird’s-Eye View Features for 3D Object Detection. IEEE Trans. Instrum. Meas. 2023, 72, 8503814. [Google Scholar] [CrossRef]
  39. Xiong, W.; Liu, J.; Huang, T.; Han, Q.L.; Xia, Y.; Zhu, B. Lxl: Lidar exclusive lean 3d object detection with 4d imaging radar and camera fusion. arXiv 2023, arXiv:2307.00724. [Google Scholar] [CrossRef]
  40. Xie, T.; Wang, S.; Wang, K.; Yang, L.; Jiang, Z.; Zhang, X.; Dai, K.; Li, R.; Cheng, J. Poly-PC: A Polyhedral Network for Multiple Point Cloud Tasks at Once. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 1233–1243. [Google Scholar]
  41. Xie, T.; Wang, K.; Lu, S.; Zhang, Y.; Dai, K.; Li, X.; Xu, J.; Wang, L.; Zhao, L.; Zhang, X.; et al. CO-Net: Learning Multiple Point Cloud Tasks at Once with a Cohesive Network. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 3523–3533. [Google Scholar]
  42. Xie, T.; Wang, L.; Wang, K.; Li, R.; Zhang, X.; Zhang, H.; Yang, L.; Liu, H.; Li, J. FARP-Net: Local-Global Feature Aggregation and Relation-Aware Proposals for 3D Object Detection. IEEE Trans. Multimed. 2023, 1–15. [Google Scholar] [CrossRef]
  43. Wang, L.; Xie, T.; Zhang, X.; Jiang, Z.; Yang, L.; Zhang, H.; Li, X.; Ren, Y.; Yu, H.; Li, J.; et al. Auto-Points: Automatic Learning for Point Cloud Analysis with Neural Architecture Search. IEEE Trans. Multimed. 2023, 1–16. [Google Scholar] [CrossRef]
  44. Palffy, A.; Pool, E.; Baratam, S.; Kooij, J.F.; Gavrila, D.M. Multi-class road user detection with 3+ 1D radar in the View-of-Delft dataset. IEEE Robot. Autom. Lett. 2022, 7, 4961–4968. [Google Scholar] [CrossRef]
  45. Team, O.; et al. Openpcdet: An open-source toolbox for 3d object detection from point clouds. OD Team 2020. Available online: https://github.com/open-mmlab/OpenPCDet (accessed on 22 October 2023).
Figure 1. Point cloud visualization collected by different sensors. (a) LiDAR. (b) 4D Radar single frame. (c) 4D Radar 5 frames.
Figure 1. Point cloud visualization collected by different sensors. (a) LiDAR. (b) 4D Radar single frame. (c) 4D Radar 5 frames.
Sensors 23 09429 g001
Figure 2. The overall architecture of our proposed SMIFormer. Our approach employs a 4D Imaging Radar point cloud as input for 3D object detection. To enable multi-view feature encoding and interaction, we introduce a parallel fusion model. Decoupled feature maps are colored by view directions with blue for bird’s-eye view, red for side view and green for front view.
Figure 2. The overall architecture of our proposed SMIFormer. Our approach employs a 4D Imaging Radar point cloud as input for 3D object detection. To enable multi-view feature encoding and interaction, we introduce a parallel fusion model. Decoupled feature maps are colored by view directions with blue for bird’s-eye view, red for side view and green for front view.
Sensors 23 09429 g002
Figure 3. Voxel feature query module. Given a specific voxel location, we project its 3D coordinates onto each of the three axially aligned planes (BEV, SV and FV). Afterwards, we extract the features of each projection from a sparse 2D feature map and consolidate the resulting features into a single feature vector.
Figure 3. Voxel feature query module. Given a specific voxel location, we project its 3D coordinates onto each of the three axially aligned planes (BEV, SV and FV). Afterwards, we extract the features of each projection from a sparse 2D feature map and consolidate the resulting features into a single feature vector.
Sensors 23 09429 g003
Figure 4. Visualization results on the VoD validation set. Each row contains two sets of data frames of images and radar points (gray points). The left side is the ground truth, the right side is the prediction result, the red box represents the car, the blue box represents the cyclist, and the green box represents the pedestrian. The local region of the point cloud, scaled by the orange wire box, demonstrates that for a few undetected objects, there are few to no points in these ground truth bounding boxes.
Figure 4. Visualization results on the VoD validation set. Each row contains two sets of data frames of images and radar points (gray points). The left side is the ground truth, the right side is the prediction result, the red box represents the car, the blue box represents the cyclist, and the green box represents the pedestrian. The local region of the point cloud, scaled by the orange wire box, demonstrates that for a few undetected objects, there are few to no points in these ground truth bounding boxes.
Sensors 23 09429 g004
Table 1. Comparison on the evaluation set of VoD dataset. All the approaches are executed with 5-scans radar detection points.
Table 1. Comparison on the evaluation set of VoD dataset. All the approaches are executed with 5-scans radar detection points.
MethodModalityEntire Annotation AreaDriving Corridor Area
CarPedCycmAPCarPedCycmAP
PointPillars [29]R37.2432.1966.8045.4170.5543.2888.1367.32
SECOND [22]R40.4030.6462.5144.5272.2541.1983.3965.61
CenterPoint [2]R32.7438.0065.5145.4262.0148.1884.9865.06
LXL-R [39]R32.7539.6568.1346.8470.2647.3487.9368.51
RCFusion [38]R + C41.7038.9568.3149.6571.8747.5088.3369.23
OursR39.5341.8864.9148.7777.0453.4082.9571.13
Table 2. Module’s ablation study on the VoD validation dataset. All the approaches are executed with 5-scans radar detection points on VoD dataset.
Table 2. Module’s ablation study on the VoD validation dataset. All the approaches are executed with 5-scans radar detection points on VoD dataset.
#MVDSAICAOEntire Annotation AreaDriving Corridor Area
CarPedCycmAPCarPedCycmAP
1 37.8239.4462.1846.4872.0849.2983.7168.36
2 41.1640.0761.0647.4378.7352.4374.8068.65
3 39.3239.9263.5347.5970.6453.7985.5770.00
439.5341.8864.9148.7777.0453.4082.9571.13
Table 3. Splitting window size ablation study on the VoD validation dataset. All the approaches are executed with 5-scans radar detection points on VoD dataset.
Table 3. Splitting window size ablation study on the VoD validation dataset. All the approaches are executed with 5-scans radar detection points on VoD dataset.
Window SizeEntire Annotation AreaDriving Corridor Area
CarPedCycmAPCarPedCycmAP
243.4838.5963.1848.4179.5951.1079.9570.21
439.5341.8864.9148.7777.0453.4082.9571.13
839.3535.6964.4946.5177.5847.9886.3770.64
Table 4. Latency evaluation for different splitting window sizes on a single GPU.
Table 4. Latency evaluation for different splitting window sizes on a single GPU.
Window SizeLatency
268 ms
461 ms
865 ms
3269 ms
128 (w/o splitting)90 ms
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shi, W.; Zhu, Z.; Zhang, K.; Chen, H.; Yu, Z.; Zhu, Y. SMIFormer: Learning Spatial Feature Representation for 3D Object Detection from 4D Imaging Radar via Multi-View Interactive Transformers. Sensors 2023, 23, 9429. https://doi.org/10.3390/s23239429

AMA Style

Shi W, Zhu Z, Zhang K, Chen H, Yu Z, Zhu Y. SMIFormer: Learning Spatial Feature Representation for 3D Object Detection from 4D Imaging Radar via Multi-View Interactive Transformers. Sensors. 2023; 23(23):9429. https://doi.org/10.3390/s23239429

Chicago/Turabian Style

Shi, Weigang, Ziming Zhu, Kezhi Zhang, Huanlei Chen, Zhuoping Yu, and Yu Zhu. 2023. "SMIFormer: Learning Spatial Feature Representation for 3D Object Detection from 4D Imaging Radar via Multi-View Interactive Transformers" Sensors 23, no. 23: 9429. https://doi.org/10.3390/s23239429

APA Style

Shi, W., Zhu, Z., Zhang, K., Chen, H., Yu, Z., & Zhu, Y. (2023). SMIFormer: Learning Spatial Feature Representation for 3D Object Detection from 4D Imaging Radar via Multi-View Interactive Transformers. Sensors, 23(23), 9429. https://doi.org/10.3390/s23239429

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop