0% found this document useful (0 votes)
46 views

Tj4Dradset: A 4D Radar Dataset For Autonomous Driving

The document introduces a new 4D radar dataset called TJ4DRadSet for autonomous driving research. It was collected using a platform with 4D radar, camera, lidar and GNSS sensors across various driving scenarios. The dataset contains over 7700 synchronized frames with 3D object annotations that can be used to develop 4D radar based perception algorithms and evaluate their performance.

Uploaded by

fake
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
46 views

Tj4Dradset: A 4D Radar Dataset For Autonomous Driving

The document introduces a new 4D radar dataset called TJ4DRadSet for autonomous driving research. It was collected using a platform with 4D radar, camera, lidar and GNSS sensors across various driving scenarios. The dataset contains over 7700 synchronized frames with 3D object annotations that can be used to develop 4D radar based perception algorithms and evaluate their performance.

Uploaded by

fake
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 6

TJ4DRadSet: A 4D Radar Dataset for Autonomous Driving

Lianqing Zheng1, Zhixiong Ma1,*, Xichan Zhu1, Bin Tan1, Sen Li1, Kai Long1, Weiqi Sun1, Sihan
Chen1, Lu Zhang1, Mengyue Wan1, Libo Huang2, Jie Bai2,*

Abstract— The next-generation high-resolution automotive


radar (4D radar) can provide additional elevation measurement
and denser point clouds, which has great potential for 3D
sensing in autonomous driving. In this paper, we introduce a
dataset named TJ4DRadSet with 4D radar points for
autonomous driving research. The dataset was collected in
various driving scenarios, with a total of 7757 synchronized
frames in 44 consecutive sequences, which are well annotated
with 3D bounding boxes and track ids. We provide a 4D
radar-based 3D object detection baseline for our dataset to
demonstrate the effectiveness of deep learning methods for 4D
radar point clouds. The dataset can be accessed via the following
link: https://github.com/TJRadarLab/TJ4DRadSet.
I. INTRODUCTION
Autonomous driving technology [1] has recently received
much attention. The high-level autonomous driving system Figure 1. Data acquisition platform and coordinate system.
mainly consists of modules such as environment perception,
road planning, and decision execution [2]. A highly reliable, neural networks have been applied to 3D perception [6].
low-cost, high-resolution perception module is necessary for Training a 3D object detection network requires large-scale
self-driving vehicles. At the current stage, the perception data, which should cover many diverse and complex
module mainly uses sensors such as cameras, lidars, and conditions. What’s more, the ground truth of data needs to be
automotive radars to obtain environmental information of accurate for supervised learning to ensure the trained network
different modes [3]. Undeniably, the camera and lidar are is valid. Compared with camera and lidar, few autonomous
vulnerable to harsh conditions such as rain, fog, and intense driving datasets contain 4D radar, which limits the research
light, whose performance will decline significantly with the and application of deep learning in the 4D radar point cloud.
increase of adversity. In contrast, the automotive radar is To fill this gap, we proposed a 4D radar dataset for
essential because of its strong robustness [4] and autonomous driving called TJ4DRadSet. The data collection
cost-effectiveness. Due to low azimuthal resolution, platform contains multi-sensors, including 4D radar, camera,
conventional automotive radar is only used for blind-spot lidar, and Global Navigation Satellite System (GNSS), as
detection, collision warning, and other driving assistance shown in Figure 1. We hope the dataset will facilitate the
applications. The emergence of new-generation 4D radar [5] research of 4D radar-based perception algorithms. Our
makes up for the low definition of conventional automotive contributions are listed as follows:
radar and provides elevation measurement, which is well  We present a dataset named TJ4DRadSet, an
suited for applications in high-level autonomous driving. The autonomous driving dataset containing 4D radar point
four dimensions of 4D radar are range, azimuth, elevation, clouds in continuous sequences with 3D annotations,
and Doppler velocity. It also provides some other low-level which also provides multi-modal complete
features such as radar-cross-section (RCS) or signal-to-noise information on lidar, camera and GNSS.
ratio (SNR).
 TJ4DRadSet contains 40K frames of synchronized
3D object detection and tracking are essential to data, where 7757 frames, 44 sequences with
environment perception. With the development of deep high-quality annotated 3D bounding boxes and track
learning and artificial intelligence, an enormous amount of ids. The 3D annotation system uses joint multi-sensor
annotation and multi-round manual checks.
1
Lianqing Zheng, Bin Tan, Sen Li, {zhenglianqing, tanbin,  TJ4DRadSet covers various road conditions, such as
lisen}@tongji.edu.cn elevated roads, complex intersections, one-way roads,
and urban roads. It also includes bad lighting
conditions such as intense light and darkness. The
dataset is suitable for developing 3D perception
algorithms based on the 4D radar to facilitate its
application in high-level autonomous driving.
TABLE I. CURRENT DRIVING RADAR DATASETS
4D Radar Object Object 3D
Dataset Size Radar Type Other Modalities
Point Cloud Detection Tracking Annotations
nuScenes[9] Large Low Resolution Lidar&Camera    

RADIATE[12] Middle Scanning Lidar&Camera    

MulRan[13] Middle Scanning Lidar    

RadarScenes[10] Large Low Resolution Camera    

RadarRobotCar[11] Large Scanning Lidar&Camera    

Astyx[14] Small High Resolution Lidar&Camera    

RADIal[15] Middle High Resolution Lidar&Camera    

VoD[16] Middle High Resolution Lidar&Camera    

TJ4DRadSet(ours) Middle High Resolution Lidar&Camera    

 Based on TJ4DRadSet, we provide a baseline for 4D novel automotive dataset, which is the work of the same period
radar-based 3D object detection. The results show that as ours. VoD contains 8600 frames of synchronized and
4D radar has a promising potential for high-level calibrated lidar, camera, and 4D radar data acquired in urban
autonomous driving. traffic, which also provides 3D annotations and track ids.
Compared to the VoD dataset, our dataset contains much
The paper is organized as follows: Section II introduces richer and more challenging driving scenario clips.
related work on other datasets. Section III describes our
dataset in detail. In Section IV, we perform the baseline result
III. THE TJ4DRADSET DATASET
of 3D object detection based on 4D radar. A brief conclusion
and future work are presented in Section Ⅴ. In this section, we introduce sensor parameters, calibration,
data collection and annotation, then provide statistical analysis
and visualization.
II. RELATED WORK
Deep learning technique is playing an increasing role in A. Sensors
autonomous driving. It relies on a large amount of high-quality The TJ4DRadSet mainly contains 4D radar, lidar and
data. Therefore, a growing number of open dataset camera. As shown in Figure 1, the camera and lidar are
benchmarks have appeared in recent years, such as KITTI [7] mounted on the roof bracket, and the 4D radar is installed in
and Waymo Open [8], which have contributed to the the middle of the front ventilation gride. The lidar can do
advancement of autonomous driving technology. With these 360-degree scanning of environmental information, while the
benchmarks, we can evaluate the performance of different camera and 4D radar capture the information in the field of
algorithms for various tasks. view (FOV) ahead, covering the forward driving view. The
Automotive radar has proven to be an effective sensor due main parameters of each sensor are shown in TABLE Ⅱ. In
to its robustness in all weather and low price. However, many addition, the GNSS information is included and corrected by
datasets do not contain radar sensors, which limits the real-time kinematic (RTK) to achieve high-precision
application of data-driven algorithms based on radar data. positioning, which has the speed and location information of
Since the nuScenes [9] dataset was released, some datasets the ego vehicle.
with radar data started to appear, which has aroused people's
interest in radar. The comparison of each dataset containing B. Sensor Calibration
radar data is shown in TABLE Ⅰ. Some datasets contain Multi-sensor calibration is the basis for perception
low-resolution FMCW radars, such as nuScenes, and algorithms. The process mainly consists of intrinsic
RadarScenes [10], whose radar point clouds lack elevation parameters calibration, extrinsic calibration, and temporal
information for accurate 3D perception. Some datasets use alignment. The intrinsic parameters and distortion coefficients
scanning radar to collect data, such as RadarRobotCar [11], of the camera are calibrated by MATLAB Toolkit [17] and a
RADIATE [12], and MulRan [13], whose radar data are checkerboard. The distortion coefficients are used for
mainly interpreted as image data and lack Doppler velocity. correction to obtain rectified images. The intrinsic parameters
For the new-generation 4D imaging radar, the 4D point cloud of 4D radar and lidar have been calibrated offline at the
will be the primary output format, containing spatial and factory.
velocity information. Currently, Astyx [14], RADIal [15] and
VoD[16] dataset have high-resolution 4D radar sensor. Astyx It can be divided into two processes for extrinsic
has only 545 frames of point cloud data, which is small and parameters: camera and lidar extrinsic calibration; 4D radar
lacks tracking information. RADIal contains complete radar and lidar extrinsic calibration. The extrinsic parameters of the
formats, such as range-Doppler maps and point clouds, which camera and 4D radar can be obtained by performing matrix
only has 2D labeled boxes and a “Car” label. VoD dataset is a operations on the remaining two extrinsic parameters. The
TABLE II. SPECIFICATION OF THE TJ4DRADSET’S SENSOR SUITES
Parameters Resolution FOV
FPS
Sensors Range Azimuth Elevation Range Azimuth Elevation
Camera 1280px 960px 66.5° 94° 30
Lidar 0.03m 0.1°-0.4° 0.33° 120m 360° 40° 10
4D Radar 0.86m <1° <1° 400m 113° 45° 15

types, such as urban roads, elevated roads, and industrial zones.


There are complex scenarios such as object-dense
intersections and simple scenarios such as one-way streets
with a few objects. Our acquisition system is based on ROS
and all sensor data are recorded in “rosbag” completely.
The ground truth annotation of the dataset mainly includes
a 3D bounding box, class and track id for each object. The
lidar sensor has a higher point cloud density than 4D radar,
which provides a more detailed description of objects’ shapes.
(a) 4D radar point clouds (b) Lidar point clouds Therefore, our annotation system mainly relies on lidar point
clouds and images for joint annotation. However, some
Figure 2. 4D radar and lidar point clouds projection.
objects that have few lidar points due to occlusion may still
appear in 4D radar FOV because of the multipath effect, and
extrinsic parameters between the different sensors are we still label them. We finished the annotation manually and
represented as translation and rotation matrix. For camera and reviewed many rounds to ensure the quality of the dataset.
lidar extrinsic calibration, we used a checkerboard to perform
2D-3D alignment of the point cloud and image data to The 3D bounding box of each object includes the center
complete a rough calibration. Then, we manually fine-tune the point ( x, y, z ) , length, width, height (l , w, h) , and orientation
extrinsic parameters by static objects such as trees and poles in angle (yaw). In addition, we provide occlusion and truncation
the environment. For 4D radar and lidar extrinsic calibration, indicators to distinguish different difficulty levels. The dataset
we consider it as 3D-3D point cloud alignment in space. has eight classes (Car, Bus, Truck, Engineering Vehicle,
Firstly, the distance between the two sensors is measured as a Pedestrian, Motorcyclist, Cyclist, and Tricyclist). In order to
rough translation parameter. Then extrinsic parameters are have a balanced label distribution and to improve the
fine-tuned by using multiple angular reflectors in space. performance of networks, we map the “Bus” and “Engineering
Vehicle” (large) to “Truck”, the “Motorcyclist” to “Cyclist”.
All sensors work under the ROS driver. Since each sensor The class of other objects is mapped to “Other Vehicle”. The
runs at different frame rates, we align the data by using the original classes are retained so that the mapping can be
arrival time of the data as the timestamp. The final 4D radar customized as needed. We assign a unique id to each object for
and lidar point clouds are projected into the image, as shown in the tracking task. Finally, 40K frames of synchronized data are
Figure 2. extracted, of which 7757 frames in 44 consecutive sequences
are labeled.
C. Data Collection and Annotation
TJ4DRadSet was collected in Suzhou, China, in the fourth D. Dataset Statistics
quarter of 2021. Figure 3 records the location of the data
In this part, we perform some statistical analysis of the
collection. The dataset covers a wide range of driving
conditions, including various lighting conditions, such as dataset. Figure 4(a) shows the number of objects for each
normal lighting, bright light and darkness, and different road class, with “Car” being the most numerous, followed by
“Cyclist”. The amount of “Truck” and “Pedestrian” is
approximately the same. Figure 4(b) shows the speed
distribution of the ego vehicle. The distribution of the point
cloud density of lidar and 4D radar is shown in Figure 4(c)(d).
It can be seen that the 4D radar point cloud is sparser than the
laser point cloud, but radar points contain more features, such
as Doppler velocity. In addition, we also count the
distribution of the distances and orientations of the main
classes, shown in Figure 5. Some typical scenarios are
visualized in Figure 6.

IV. BASELINE EXPERIMENTS


This section establishes baselines of 3D object detection
based on 4D radar and lidar. We divide the dataset into a
Figure 3. Locations of the data collection. training set and test set by sequence and keep the test set with
(a) Number of objects (b) Ego-vehicle velocity (c) Lidar point density (d) 4D radar point density

Figure 4. Some basic statistics.

Figure 5. Distribution of the distances and orientations of “Car”, “Truck”, “Pedestrian” and “Cyclist”.

TABLE III. BASELINE RESULTS (4D RADAR)

BEV-50m 3D-50m BEV-70m 3D-70m


Class
[email protected] [email protected] [email protected] [email protected] [email protected] [email protected] [email protected] [email protected]
Car 23.06 36.73 12.63 27.96 26.19 40.14 16.85 33.30

Truck 16.76 36.37 12.64 31.33 13.46 30.49 10.07 25.51

Pedestrian 35.24 27.64 35.26 27.19

Cyclist 21.62 40.26 18.34 38.42 21.38 39.80 17.70 38.20

TABLE IV. BASELINE RESULTS (LIDAR)

BEV-50m 3D-50m BEV-70m 3D-70m


Class
[email protected] [email protected] [email protected] [email protected] [email protected] [email protected] [email protected] [email protected]
Car 69.69 69.79 69.47 69.72 52.76 52.77 52.67 52.76

Truck 37.45 49.06 30.47 43.34 24.72 33.42 22.74 31.89

Pedestrian 56.28 56.09 49.23 49.07

Cyclist 53.00 54.12 49.28 54.12 44.72 48.32 43.09 48.59

good coverage. In this way, we get 5717 training and 2040 test system through the lidar-radar extrinsic matrix. Due to the
samples and keep the data split fixed. sparsity of radar point clouds, some of the existing networks
are difficult to be applied directly to this data format. In this
The original annotations are under the lidar coordinate paper, we use PointPillars [18] as the baseline algorithm for
system, and we transfer the labels to the 4D radar coordinate
Figure 6. Visualization of typical samples of TJ4DRadSet.
4D radar and lidar because of its good adaptability and the [3] R. Ravindran, M. J. Santora and M. M. Jamali, “Multi-Object Detection
trade-off in both speed and precision. To adapt the 4D radar and Tracking, Based on DNN, for Autonomous Vehicles: A Review,”
IEEE Sensors Journal, vol. 21, no. 5, pp. 5668-5677, Mar. 2021.
data, we have partially modified the original configuration and [4] S. Lee, Y. J. Yoon, J. E. Lee and S. C. Kim, “Human–vehicle
retrained the model using TJ4DRadSet. The detection range classification using feature-based SVM in 77-GHz automotive FMCW
along the x-axis is set to 69.12m. We use five-dimensional radar,” IET Radar, Sonar and Navigation, vol. 11, no. 10, pp.
features of radar point clouds, which include spatial 1589-1596, Aug. 2017.
information ( x, y, z ) , Doppler velocity (v) and signal to noise [5] S. Brisken, F.Ruf, F. Höhne, “Recent evolution of automotive imaging
radar and its information content,” IET Radar, Sonar and Navigation,
ratio ( s ) . The Doppler velocity (v) is the absolute radial vol. 12, no. 10, pp. 1078-1081, Oct. 2018.
velocity after compensation by ego-motion. In terms of [6] E. Arnold, O. Y. Al-Jarrah, M. Dianati, S. Fallah, D. Oxtoby and A.
Mouzakitis, “A Survey on 3D Object Detection Methods for
network parameters, we choose the pillar size to be Autonomous Driving Applications,” IEEE Transactions on Intelligent
(0.16m, 0.16m) . The anchor size format is defined as (l , w, h) . Transportation Systems, vol. 20, no. 10, pp. 3782-3795, Oct. 2019.
For the four classes (“Car”, “Truck”, “Pedestrian”, and [7] A. Geiger, P . Lenz, C. Stiller, and R. Urtasun, “Vision meets
robotics:The KITTI dataset,” International Journal of Robotics
“Cyclist”), the anchor sizes are listed as Research, vol. 32, no. 11, pp. 1231-1237, Sep. 2013.
follows: (4.56m,1.84m,1.70m) , (10.76m, 2.66m,3.47 m) , [8] P . Sun, H. Kretzschmar, X. Dotiwalla, A. Chouard, V . Patnaik, P .
(0.80m, 0.60m,1.69m) , (1.77 m, 0.78m,1.60m) . Besides, Tsui, J. Guo, Y . Zhou, Y . Chai, B. Caineet al., “Scalability in
perception for autonomous driving: Waymo open dataset,” in
some data augmentations are used to enhance the robustness of Proceedings of the IEEE/CVF Conference on Computer Vision and
the network, including the world random rotation and random Pattern Recognition (CVPR), 2020, pp. 2446–2454.
scaling. We use the Adam optimizer [19] to train the model for [9] H. Caesar, V . Bankiti, A. H. Lang, S. V ora, V . E. Liong, Q. Xu, A.
Krishnan, Y . Pan, G. Baldan, and O. Beijbom, “nuScenes: A
80 epochs.
multimodal dataset for autonomous driving,” in Proceedings of the
In the evaluation stage, the average precision (AP) is chosen IEEE/CVF Conference on Computer Vision and Pattern Recognition
(CVPR), 2020, pp. 11621–11631.
as the metric to evaluate the detection results for each class. [10] O. Schumann et al., “RadarScenes: A Real-World Radar Point Cloud
Specifically, we use 0.5 and 0.25 IoU thresholds to test “Car”, Data Set for Automotive Applications,” in 2021 IEEE 24th
“Truck”, and “Cyclist”, and only use the 0.25 IoU threshold to International Conference on Information Fusion (FUSION), 2021, pp.
evaluate “Pedestrian”. We denote the AP under these two 1-8.
thresholds as [email protected] and [email protected]. TABLE Ⅲ and TABLE [11] D. Barnes, M. Gadd, P. Murcutt, P. Newman and I. Posner, “The
Ⅳ show the baseline performance at different distances (50m Oxford Radar RobotCar Dataset: A Radar Extension to the Oxford
RobotCar Dataset,”in 2020 IEEE International Conference on Robotics
and 70m) and views(BEV, 3D) using 4D radar and lidar, and Automation (ICRA), 2020, pp. 6433-6438.
respectively. The results clearly illustrate that 4D radar has [12] M. Sheeny, E. De Pellegrin, S. Mukherjee, A. Ahrabian, S. Wang and A.
potential for 3D perception. In the BEV view, the average Wallace, “RADIATE: A Radar Dataset for Automotive Perception in
accuracy for all classes is over 30% at the 0.25 IoU threshold. Bad Weather,” in 2021 IEEE International Conference on Robotics and
Although the baseline algorithm can achieve some results, Automation (ICRA), 2021, pp. 1-7.
[13] G. Kim, Y. S. Park, Y. Cho, J. Jeong and A. Kim, “MulRan:
there is still a big gap between 4D radar and lidar. Under the Multimodal Range Dataset for Urban Place Recognition,” in 2020
same algorithm (PointPillars), lidar detection results IEEE International Conference on Robotics and Automation (ICRA),
completely outperform 4D radar, which can be due to several 2020, pp. 6246-6253.
reasons. First, 4D radar has a lower point density, which [14] M. Meyer and G. Kuschk, “Automotive Radar Dataset for Deep
maybe makes it difficult for the baseline network to extract Learning Based 3D Object Detection,” in 2019 16th European Radar
Conference (EuRAD), 2019, pp. 129-132.
features effectively. In addition, different data augmentations [15] J. Rebut, A. Ouaknine, W. Malik, P. Pérez, “Raw High-Definition
could also have an impact on the results. It is of great concern Radar for Multi-Task Learning,” arXiv preprint arXiv: 2112.10646,
how to better extract 4D radar point cloud features and fuse 2021.
information from other modalities. [16] A. Palffy, E. Pool, S. Baratam, J. F. P. Kooij, and D. M. Gavrila,
“Multi-Class Road User Detection With 3+1D Radar in the
View-of-Delft Dataset.” IEEE Robotics and Automation Letters, vol. 7,
V. CONCLUSION AND FUTURE WORK no. 2, pp. 4961-4968, April. 2022.
[17] “Matlab camera calibration,” https://www.mathworks.com/help/vision/
In this paper, we introduce TJ4DRadSet, a multi-modal ug/camera-calibration.html, 2021.
autonomous driving dataset containing 4D radar point cloud. [18] A. H. Lang, S. V ora, H. Caesar, L. Zhou, J. Yang, and O. Beijbom,
“Pointpillars: Fast encoders for object detection from point clouds,” in
The dataset is used to study 4D radar-based 3D perception Proceedings of the IEEE/CVF Conference on Computer Vision and
algorithms. We provide a detailed description of the dataset Pattern Recognition (CVPR), 2019, pp. 12697–12705.
and conduct baseline experiments. In the future, we will [19] D. P. Kingma, J. Ba, “Adam: A Method for Stochastic Optimization”,
arXiv preprint arXiv: 1412.6980, 2014.
further expand the dataset and research fusion algorithms,
point cloud enhancement and feature representation based on
4D radar.

REFERENCES
[1] A. Agafonov and A. Yumaganov, “3D Objects Detection in an
Autonomous Car Driving Problem,” in 2020 International Conference
on Information Technology and Nanotechnology (ITNT), 2020, pp. 1-5.
[2] X. Duan, H. Jiang, D. Tian, T. Zou, J. Zhou and Y. Cao, “V2I based
environment perception for autonomous vehicles at intersections,”
China Communications, vol. 18, no. 7, pp. 1-12, July 2021.

You might also like