Tj4Dradset: A 4D Radar Dataset For Autonomous Driving
Tj4Dradset: A 4D Radar Dataset For Autonomous Driving
Lianqing Zheng1, Zhixiong Ma1,*, Xichan Zhu1, Bin Tan1, Sen Li1, Kai Long1, Weiqi Sun1, Sihan
Chen1, Lu Zhang1, Mengyue Wan1, Libo Huang2, Jie Bai2,*
Based on TJ4DRadSet, we provide a baseline for 4D novel automotive dataset, which is the work of the same period
radar-based 3D object detection. The results show that as ours. VoD contains 8600 frames of synchronized and
4D radar has a promising potential for high-level calibrated lidar, camera, and 4D radar data acquired in urban
autonomous driving. traffic, which also provides 3D annotations and track ids.
Compared to the VoD dataset, our dataset contains much
The paper is organized as follows: Section II introduces richer and more challenging driving scenario clips.
related work on other datasets. Section III describes our
dataset in detail. In Section IV, we perform the baseline result
III. THE TJ4DRADSET DATASET
of 3D object detection based on 4D radar. A brief conclusion
and future work are presented in Section Ⅴ. In this section, we introduce sensor parameters, calibration,
data collection and annotation, then provide statistical analysis
and visualization.
II. RELATED WORK
Deep learning technique is playing an increasing role in A. Sensors
autonomous driving. It relies on a large amount of high-quality The TJ4DRadSet mainly contains 4D radar, lidar and
data. Therefore, a growing number of open dataset camera. As shown in Figure 1, the camera and lidar are
benchmarks have appeared in recent years, such as KITTI [7] mounted on the roof bracket, and the 4D radar is installed in
and Waymo Open [8], which have contributed to the the middle of the front ventilation gride. The lidar can do
advancement of autonomous driving technology. With these 360-degree scanning of environmental information, while the
benchmarks, we can evaluate the performance of different camera and 4D radar capture the information in the field of
algorithms for various tasks. view (FOV) ahead, covering the forward driving view. The
Automotive radar has proven to be an effective sensor due main parameters of each sensor are shown in TABLE Ⅱ. In
to its robustness in all weather and low price. However, many addition, the GNSS information is included and corrected by
datasets do not contain radar sensors, which limits the real-time kinematic (RTK) to achieve high-precision
application of data-driven algorithms based on radar data. positioning, which has the speed and location information of
Since the nuScenes [9] dataset was released, some datasets the ego vehicle.
with radar data started to appear, which has aroused people's
interest in radar. The comparison of each dataset containing B. Sensor Calibration
radar data is shown in TABLE Ⅰ. Some datasets contain Multi-sensor calibration is the basis for perception
low-resolution FMCW radars, such as nuScenes, and algorithms. The process mainly consists of intrinsic
RadarScenes [10], whose radar point clouds lack elevation parameters calibration, extrinsic calibration, and temporal
information for accurate 3D perception. Some datasets use alignment. The intrinsic parameters and distortion coefficients
scanning radar to collect data, such as RadarRobotCar [11], of the camera are calibrated by MATLAB Toolkit [17] and a
RADIATE [12], and MulRan [13], whose radar data are checkerboard. The distortion coefficients are used for
mainly interpreted as image data and lack Doppler velocity. correction to obtain rectified images. The intrinsic parameters
For the new-generation 4D imaging radar, the 4D point cloud of 4D radar and lidar have been calibrated offline at the
will be the primary output format, containing spatial and factory.
velocity information. Currently, Astyx [14], RADIal [15] and
VoD[16] dataset have high-resolution 4D radar sensor. Astyx It can be divided into two processes for extrinsic
has only 545 frames of point cloud data, which is small and parameters: camera and lidar extrinsic calibration; 4D radar
lacks tracking information. RADIal contains complete radar and lidar extrinsic calibration. The extrinsic parameters of the
formats, such as range-Doppler maps and point clouds, which camera and 4D radar can be obtained by performing matrix
only has 2D labeled boxes and a “Car” label. VoD dataset is a operations on the remaining two extrinsic parameters. The
TABLE II. SPECIFICATION OF THE TJ4DRADSET’S SENSOR SUITES
Parameters Resolution FOV
FPS
Sensors Range Azimuth Elevation Range Azimuth Elevation
Camera 1280px 960px 66.5° 94° 30
Lidar 0.03m 0.1°-0.4° 0.33° 120m 360° 40° 10
4D Radar 0.86m <1° <1° 400m 113° 45° 15
Figure 5. Distribution of the distances and orientations of “Car”, “Truck”, “Pedestrian” and “Cyclist”.
good coverage. In this way, we get 5717 training and 2040 test system through the lidar-radar extrinsic matrix. Due to the
samples and keep the data split fixed. sparsity of radar point clouds, some of the existing networks
are difficult to be applied directly to this data format. In this
The original annotations are under the lidar coordinate paper, we use PointPillars [18] as the baseline algorithm for
system, and we transfer the labels to the 4D radar coordinate
Figure 6. Visualization of typical samples of TJ4DRadSet.
4D radar and lidar because of its good adaptability and the [3] R. Ravindran, M. J. Santora and M. M. Jamali, “Multi-Object Detection
trade-off in both speed and precision. To adapt the 4D radar and Tracking, Based on DNN, for Autonomous Vehicles: A Review,”
IEEE Sensors Journal, vol. 21, no. 5, pp. 5668-5677, Mar. 2021.
data, we have partially modified the original configuration and [4] S. Lee, Y. J. Yoon, J. E. Lee and S. C. Kim, “Human–vehicle
retrained the model using TJ4DRadSet. The detection range classification using feature-based SVM in 77-GHz automotive FMCW
along the x-axis is set to 69.12m. We use five-dimensional radar,” IET Radar, Sonar and Navigation, vol. 11, no. 10, pp.
features of radar point clouds, which include spatial 1589-1596, Aug. 2017.
information ( x, y, z ) , Doppler velocity (v) and signal to noise [5] S. Brisken, F.Ruf, F. Höhne, “Recent evolution of automotive imaging
radar and its information content,” IET Radar, Sonar and Navigation,
ratio ( s ) . The Doppler velocity (v) is the absolute radial vol. 12, no. 10, pp. 1078-1081, Oct. 2018.
velocity after compensation by ego-motion. In terms of [6] E. Arnold, O. Y. Al-Jarrah, M. Dianati, S. Fallah, D. Oxtoby and A.
Mouzakitis, “A Survey on 3D Object Detection Methods for
network parameters, we choose the pillar size to be Autonomous Driving Applications,” IEEE Transactions on Intelligent
(0.16m, 0.16m) . The anchor size format is defined as (l , w, h) . Transportation Systems, vol. 20, no. 10, pp. 3782-3795, Oct. 2019.
For the four classes (“Car”, “Truck”, “Pedestrian”, and [7] A. Geiger, P . Lenz, C. Stiller, and R. Urtasun, “Vision meets
robotics:The KITTI dataset,” International Journal of Robotics
“Cyclist”), the anchor sizes are listed as Research, vol. 32, no. 11, pp. 1231-1237, Sep. 2013.
follows: (4.56m,1.84m,1.70m) , (10.76m, 2.66m,3.47 m) , [8] P . Sun, H. Kretzschmar, X. Dotiwalla, A. Chouard, V . Patnaik, P .
(0.80m, 0.60m,1.69m) , (1.77 m, 0.78m,1.60m) . Besides, Tsui, J. Guo, Y . Zhou, Y . Chai, B. Caineet al., “Scalability in
perception for autonomous driving: Waymo open dataset,” in
some data augmentations are used to enhance the robustness of Proceedings of the IEEE/CVF Conference on Computer Vision and
the network, including the world random rotation and random Pattern Recognition (CVPR), 2020, pp. 2446–2454.
scaling. We use the Adam optimizer [19] to train the model for [9] H. Caesar, V . Bankiti, A. H. Lang, S. V ora, V . E. Liong, Q. Xu, A.
Krishnan, Y . Pan, G. Baldan, and O. Beijbom, “nuScenes: A
80 epochs.
multimodal dataset for autonomous driving,” in Proceedings of the
In the evaluation stage, the average precision (AP) is chosen IEEE/CVF Conference on Computer Vision and Pattern Recognition
(CVPR), 2020, pp. 11621–11631.
as the metric to evaluate the detection results for each class. [10] O. Schumann et al., “RadarScenes: A Real-World Radar Point Cloud
Specifically, we use 0.5 and 0.25 IoU thresholds to test “Car”, Data Set for Automotive Applications,” in 2021 IEEE 24th
“Truck”, and “Cyclist”, and only use the 0.25 IoU threshold to International Conference on Information Fusion (FUSION), 2021, pp.
evaluate “Pedestrian”. We denote the AP under these two 1-8.
thresholds as [email protected] and [email protected]. TABLE Ⅲ and TABLE [11] D. Barnes, M. Gadd, P. Murcutt, P. Newman and I. Posner, “The
Ⅳ show the baseline performance at different distances (50m Oxford Radar RobotCar Dataset: A Radar Extension to the Oxford
RobotCar Dataset,”in 2020 IEEE International Conference on Robotics
and 70m) and views(BEV, 3D) using 4D radar and lidar, and Automation (ICRA), 2020, pp. 6433-6438.
respectively. The results clearly illustrate that 4D radar has [12] M. Sheeny, E. De Pellegrin, S. Mukherjee, A. Ahrabian, S. Wang and A.
potential for 3D perception. In the BEV view, the average Wallace, “RADIATE: A Radar Dataset for Automotive Perception in
accuracy for all classes is over 30% at the 0.25 IoU threshold. Bad Weather,” in 2021 IEEE International Conference on Robotics and
Although the baseline algorithm can achieve some results, Automation (ICRA), 2021, pp. 1-7.
[13] G. Kim, Y. S. Park, Y. Cho, J. Jeong and A. Kim, “MulRan:
there is still a big gap between 4D radar and lidar. Under the Multimodal Range Dataset for Urban Place Recognition,” in 2020
same algorithm (PointPillars), lidar detection results IEEE International Conference on Robotics and Automation (ICRA),
completely outperform 4D radar, which can be due to several 2020, pp. 6246-6253.
reasons. First, 4D radar has a lower point density, which [14] M. Meyer and G. Kuschk, “Automotive Radar Dataset for Deep
maybe makes it difficult for the baseline network to extract Learning Based 3D Object Detection,” in 2019 16th European Radar
Conference (EuRAD), 2019, pp. 129-132.
features effectively. In addition, different data augmentations [15] J. Rebut, A. Ouaknine, W. Malik, P. Pérez, “Raw High-Definition
could also have an impact on the results. It is of great concern Radar for Multi-Task Learning,” arXiv preprint arXiv: 2112.10646,
how to better extract 4D radar point cloud features and fuse 2021.
information from other modalities. [16] A. Palffy, E. Pool, S. Baratam, J. F. P. Kooij, and D. M. Gavrila,
“Multi-Class Road User Detection With 3+1D Radar in the
View-of-Delft Dataset.” IEEE Robotics and Automation Letters, vol. 7,
V. CONCLUSION AND FUTURE WORK no. 2, pp. 4961-4968, April. 2022.
[17] “Matlab camera calibration,” https://www.mathworks.com/help/vision/
In this paper, we introduce TJ4DRadSet, a multi-modal ug/camera-calibration.html, 2021.
autonomous driving dataset containing 4D radar point cloud. [18] A. H. Lang, S. V ora, H. Caesar, L. Zhou, J. Yang, and O. Beijbom,
“Pointpillars: Fast encoders for object detection from point clouds,” in
The dataset is used to study 4D radar-based 3D perception Proceedings of the IEEE/CVF Conference on Computer Vision and
algorithms. We provide a detailed description of the dataset Pattern Recognition (CVPR), 2019, pp. 12697–12705.
and conduct baseline experiments. In the future, we will [19] D. P. Kingma, J. Ba, “Adam: A Method for Stochastic Optimization”,
arXiv preprint arXiv: 1412.6980, 2014.
further expand the dataset and research fusion algorithms,
point cloud enhancement and feature representation based on
4D radar.
REFERENCES
[1] A. Agafonov and A. Yumaganov, “3D Objects Detection in an
Autonomous Car Driving Problem,” in 2020 International Conference
on Information Technology and Nanotechnology (ITNT), 2020, pp. 1-5.
[2] X. Duan, H. Jiang, D. Tian, T. Zou, J. Zhou and Y. Cao, “V2I based
environment perception for autonomous vehicles at intersections,”
China Communications, vol. 18, no. 7, pp. 1-12, July 2021.