Next Article in Journal
Data Mining and Fusion Framework for In-Home Monitoring Applications
Previous Article in Journal
Quantification of the Risk of Musculoskeletal Disorders of the Upper Limb Using Fuzzy Logic: A Study of Manual Wheelchair Propulsion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

L-DIG: A GAN-Based Method for LiDAR Point Cloud Processing under Snow Driving Conditions

1
Graduate School of Informatics, Nagoya University, Furo-cho, Chikusa-Ward, Nagoya 464-8601, Japan
2
Institutes of Innovation for Future Society, Nagoya University, Furo-cho, Chikusa-Ward, Nagoya 464-8601, Japan
3
Tier IV Inc., Nagoya University Open Innovation Center, 1-3, Mei-eki 1-chome, Nakamura-Ward, Nagoya 450-6610, Japan
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(21), 8660; https://doi.org/10.3390/s23218660
Submission received: 7 August 2023 / Revised: 4 October 2023 / Accepted: 16 October 2023 / Published: 24 October 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
LiDAR point clouds are significantly impacted by snow in driving scenarios, introducing scattered noise points and phantom objects, thereby compromising the perception capabilities of autonomous driving systems. Current effective methods for removing snow from point clouds largely rely on outlier filters, which mechanically eliminate isolated points. This research proposes a novel translation model for LiDAR point clouds, the ‘L-DIG’ (LiDAR depth images GAN), built upon refined generative adversarial networks (GANs). This model not only has the capacity to reduce snow noise from point clouds, but it also can artificially synthesize snow points onto clear data. The model is trained using depth image representations of point clouds derived from unpaired datasets, complemented by customized loss functions for depth images to ensure scale and structure consistencies. To amplify the efficacy of snow capture, particularly in the region surrounding the ego vehicle, we have developed a pixel-attention discriminator that operates without downsampling convolutional layers. Concurrently, the other discriminator equipped with two-step downsampling convolutional layers has been engineered to effectively handle snow clusters. This dual-discriminator approach ensures robust and comprehensive performance in tackling diverse snow conditions. The proposed model displays a superior ability to capture snow and object features within LiDAR point clouds. A 3D clustering algorithm is employed to adaptively evaluate different levels of snow conditions, including scattered snowfall and snow swirls. Experimental findings demonstrate an evident de-snowing effect, and the ability to synthesize snow effects.

1. Introduction

The presence of weather conditions affects and corrupts the signal quality of autonomous driving system (ADS) sensors and causes perception failures [1]. During the past decade, with the expansion of weather-related datasets [2,3] and the development of sophisticated machine learning techniques, models designed to address adverse weather problems, such as precipitation, in autonomous driving have been widely studied [4]. Progress has been made in the weather models of both LiDAR and images with weather conditions commonly treated as uniformly or normally distributed noise that can be represented by linear or monotone functions [5,6]. Snow is among the conditions with tangible threats to the sensors but few profundities in research. What makes it even more special is the unexpected irregular swirls phenomena brought by wind or the motions of the ego vehicle itself or passing-by vehicles [7], causing not only randomly distributed scattered noise (salt-and-pepper) but also swirl clusters in LiDAR point clouds, as shown in the disordered clusters near the center in the red boxes of Figure 1a. The study on the snow problem in the point cloud has been focusing on k-d-tree-based neighbor searching outlier filters [8] in recent years and the de-noising performance has almost reached saturation [9]. Nonetheless, few attempts at the implementation of deep-learning-based models have been made in snow conditions. Unlike filters with limited explainability, learning-based models have the potential to grasp both the surface and hidden features of snow clusters in a certain driving scene and perform snow synthesizing on top of snow de-noising.
The development of robust weather models benefits from training on paired data, i.e., a pair of weather-corrupted data and clear data with the rest of the elements identical, which are commonly obtained via artificially synthesizing realistic weather effects in previously clear driving scene images [10,11,12]. Such an approach has been proven highly effective in rain [13,14], fog [15,16], and snow [17] weather conditions in camera images, plus contaminations on the camera lens [18]. However, due to the relatively low data density, the realization of the weather effects in point clouds still largely depends on the collections in weather chambers [19,20] before the mature realization of weather data augmentation in point clouds. Although models have been successfully built for point clouds under rain and fog with additional road data [21], the low domain similarities between chambers and real roads still largely limit the generality. What is more, common experimental facilities with controllable precipitation rates across the world can hardly simulate complicated weather conditions such as dynamic snowfall [4]. Therefore, it is necessary to develop a way to work with few paired data or unpaired data.
In terms of disentangled data processing, CycleGAN [22] demonstrates a high ability in style conversion and object generation [23] based on datasets with different backgrounds and from different domains, and its implementation in weather models has been proven feasible [17,24]. In this research, we propose the ‘L-DIG’ (LiDAR depth images GAN), a GAN-based method using depth image priors for LiDAR point cloud processing under various snow conditions. The original data format of the LiDAR point cloud in the spatial dimension does not align with the plane dimension as camera images, while a depth image is able to store the third dimension information of a point’s spatial coordinates into its pixel value, hence serving as the 2D representation of the LiDAR point cloud, providing the opportunity of employing GAN-based methods [25]. Our unpaired training datasets are derived from the Canadian Adverse Driving Conditions (CADC) dataset [2], which is known for its authentic winter driving scenarios and snow diversity. The proposed model aims to perform snow de-noising with a deep-level understanding of the snow features related to the driving scene, and inversely generate fake snow conditions in clear point clouds, exploring the possibility of creating paired synthetic datasets for point clouds with snow conditions.
Furthermore, the quantitative evaluation of LiDAR point cloud processing has always been a tricky task. Researchers used to select a certain amount of samples, 100 frames for example, and manually determine if a point is a snow point, in order to calculate the precision and recall of the removal of snow noise points [8,26]. Even though straightforward, it has two downsides: For one, it consumes an unimaginable amount of time and manpower to manually annotate a whole dataset while a small portion of samples suffers the risk of bias. Secondly, the accuracy of human annotation on point clouds with over 20,000 points in each frame can be as low as 85% and not satisfied enough to support the subsequent calculations on precision and recall [27]. Given that a LiDAR point cloud fundamentally represents a distribution of points within a three-dimensional space, 3D point clustering algorithms can be effectively used to reflect the physical properties and assess the performance of point cloud processing specifically on snow points [28].
Among various algorithms, the ordering points to identify the clustering structure (OPTICS) algorithm [29] emerges as a preferred choice due to its exceptional capacity in handling clusters of varying densities [30]. Derived from the DBSCAN (density-based spatial clustering of applications with noise) methodology, OPTICS identifies distinct cluster groups by densely associating data points within a predetermined radius [31]. This principle highly aligns with the adaptive filters, which are predicated on neighbor searching and have been widely employed in the LiDAR point cloud de-noising over recent years.
An example of the OPTICS clustering result is shown in Figure 1b. It can be seen that despite the density variation, everything has been classified into separated groups of clusters, each distinguished by unique colors, while the environmental structures are also well segmented. The conglomeration of snow swirl points, positioned at the lower left of the center, are collectively assigned to large blue and purple clusters. Minor snow clusters, such as those in the immediate right vicinity of the center, along with individual scattered snow points spread across the scene, are categorized into smaller, uniquely colored clusters.
The main contributions of this research are described as follows:
  • We build a deep-learning-based LiDAR point cloud translation model with unpaired data and depth image priors. A new discriminator structure to better remove snow noise and new loss functions, including depth and SSIM (structural similarity index measure) losses to maintain the driving scene integrity have been designed in the proposed model.
  • The proposed model is able to perform LiDAR point cloud translations between snow conditions and clear conditions in driving scenes. The model demonstrates a certain level of understanding of the snow features and performs both effective snow de-noising and artificial snow point generation, which could help create paired datasets for training or simulation in autonomous driving applications.
  • We employ the OPTICS clustering algorithm as the quantitative evaluation metrics of the snow conditions in LiDAR point clouds. We set adaptive parameters to cluster different forms and levels of snow masses, calculate multiple indexes, and present distribution changes in violin plots to reflect the conditions of snow in the whole dataset in a comprehensive way.

2. Related Works

Our proposed model is a GAN-based deep-learning model for LiDAR point cloud processing including de-noising. As a result, we are going to introduce some related works regarding adaptive noise removal filters, deep-learning-based LiDAR point cloud processing methods, and GAN-based LiDAR translation methods in the following sections.

2.1. Adaptive Noise Removal Filters in LiDAR Point Cloud Processing

Noise point removal, or to say de-noising, is one of the most fundamental preliminary tasks of LiDAR point cloud processing to enhance the data quality and is commonly realized by radius outlier removal (ROR) filters. In the 3D space, ROR filters methodically analyze each individual point, assessing neighboring points located within a predefined radius. If the count of these neighboring points is below a certain threshold ( k m i n ), the filter deems the point as noise and eliminates it. This approach is notably effective in removing noise generated by small, separate solid entities such as snowflakes. However, it carries a potential drawback as it might also eliminate distant points in the environment, thus compromising the data integrity of the original point clouds. One of the main reasons for this limitation lies in the mismatch between the fixed predefined search radius and the varying point density that tends to decrease towards the edges of the point clouds. As a result, there is a compelling demand for an adaptive search radius to rectify this discrepancy.
Charron et al. [8] proposed an advanced variant of the conventional filter, referred to as the dynamic radius outlier removal (DROR) filter. The distinctive characteristic of the DROR filter is that it dynamically adjusts the search radius ( S R p ) of each point based on its intrinsic geometric attributes, as shown in (1), where r p is the range from the sensor to the point p, α is the horizontal angular resolution of the LiDAR, the product of ( r p α ) represents point spacing, and β d is the multiplication factor. This approach meticulously eliminates central noise points, while facilitating the retention of pivotal points situated at greater distances within the point clouds, successfully preserving the pertinent structures in distance.
S R p = β d ( r p α )
Another noteworthy innovation is the dynamic statistical outlier removal (DSOR) filter introduced by Kurup and Bos [26]. In this approach, point removal involves a thorough examination of each point to ascertain if the mean distance to its k nearest neighbors exceeds a dynamic threshold. The calculation of this dynamic threshold is defined as follows:
T = μ + σ β s
T d = r ( T r p )
where the variables are designated as follows: μ symbolizes the overall mean distance between every point and their corresponding k nearest neighbors. The global standard deviation of these distances is represented by σ . The term β s stands for a predetermined multiplier parameter. The distance from the sensor to a given point p is indicated by r p , whereas r signifies a scaling factor for the spacing between points. In terms of performance, DSOR proves to be either on par with or superior to DROR in both noise reduction and environmental feature preservation. Remarkably, DSOR also exhibits a significantly higher computational speed compared to DROR, further underlining its advantages in data preparation tasks.
Subsequently, a variety of optimized filters based on adaptive parameters such as intensity-based [32], and density-based [33] were put forward. While adaptive filters have made strides in de-noising, they are inherently constrained by predefined removal rules. This limitation becomes particularly evident when considering the diverse and infinite ways that snow can form clusters, a complexity that cannot be fully captured by such rigid algorithms.

2.2. Deep-Learning-Based LiDAR Point Cloud Processing

The initial implementation of point cloud de-noising using deep learning was primarily focused on mitigating the effects of rain and fog, which introduce obscurity or diffusion of LiDAR signals due to the presence of small water droplets. Additionally, external disturbances such as wind or spray can lead to the formation of clustered fog and water mist, thereby causing false obstacles for LiDAR [34]. Shamsudin et al. [35] developed algorithms for removing fog from 3D point clouds after detection using intensity and geometrical distribution to separate and target clusters of points, which were then removed from the point cloud. The problem is such algorithms are designed for indoor environments where fog behaviors are significantly different than outdoor driving scenarios.
To address the weather problems under driving conditions, Heinzler et al. [21] proposed a CNN-based approach capable of weather segmentation and de-noising. The authors collected both clear and foggy datasets within the CEREMA’s climatic chamber [20], enabling controlled precipitation rates and visibility conditions. By utilizing the paired datasets, they facilitated an automated labeling procedure that annotated weather data, enabling the model to acquire knowledge regarding the distinctive attributes of rainfall and fog reflections in point clouds. Subsequently, the trained model demonstrated the ability to discriminate fog/rain-induced point clusters and eliminate noise while preserving the original objects, such as pedestrians or cyclists. To enhance the model’s capability of comprehending real-world elements that are challenging to simulate in fog chambers, the authors also incorporated fog and rain data augmentation upon road data in the CNN model.
Besides de-noising techniques, augmentation is another valuable LiDAR point cloud processing method that can offer comparable assistance in managing adverse weather conditions. Hahner et al. [36] developed a method that simulates snow particles in a 2D space corresponding to each LiDAR line and adjusts each LiDAR beam’s measurements based on the resulting geometry. In addition, they factored in ground wetness, a common occurrence during snowfall, in their LiDAR point clouds as a supplement of the augmentation. The notable enhancement observed in the performance of 3D object detection subsequent to training on semi-synthetic snowy training data substantiates the successful simulation of snowfall. It is of particular importance to acknowledge that their snow augmentation approach predominantly focuses on light snowfall conditions under the rate of 2.5 mm/h, wherein the prevalent snow effects in LiDAR point clouds manifest as dispersed noise points rather than snow clusters. Snow clusters pose a greater challenge to LiDAR perception in actual driving conditions and our primary focus revolves around the study of snow clusters.

2.3. GAN-Based LiDAR Translation among Various Adverse Conditions

The utilization of GANs is particularly suitable for weather de-noising tasks due to the inherent challenge of obtaining diverse weather conditions while maintaining a consistent background at the pixel level, given the ever-changing atmospheric conditions. Notable methods in this area include CycleGAN [22], DiscoGAN [37], and DualGAN [38], which introduce a cycle-consistency constraint to establish connections between the images. Leveraging the capabilities of GANs to generate visually realistic images without requiring paired training data, translations between weather images and clear ones have been widely studied. De-weather frameworks based on GAN structure have proven effective in removing multiple weather conditions [24,39] in a single image.
Consequently, the adaptation of GAN models for the translation of LiDAR point clouds has emerged as a natural progression. Sallab et al. [40,41] were the first to explore the translation from simulated CARLA driving scenes to synthetic KITTI [42] point clouds, employing the CycleGAN framework. Similarly, Lee et al. [43] turned to point cloud translations between sunny, rainy, and foggy weather conditions. They employed the depth and intensity channels of the 2D Polar Grid Map (PGM) for CycleGAN processing. It is worth noting that inter-domain translations between fog chamber scenes and real-world roads present a challenge wherein artificial precipitation from sprinklers can be detected as vertical cylinders rather than genuine rainfall by LiDARs [1]. This discrepancy may impact the interpretability of weather reflection features in point clouds, thereby potentially diminishing the overall translation performance. The inability of chambers to simulate snowy conditions has also resulted in a lack of advancements in the processing of LiDAR point cloud data in snow environments.

3. Methods

3.1. LiDAR 2D Representation

To adapt LiDAR point cloud data to fit within the structure of the GAN-based model, we initially applied a dimensionality reduction to the point clouds. This yields a 2D visualization of the point clouds, namely, depth images, which signifies the orthographic projection of the point clouds. By unrolling the exterior surface of the LiDAR’s horizontal field of view (FOV) cylinder and mapping each point from the point cloud onto this frontal-view plane, a rectangular image encompassing all points within the LiDAR’s FOV is obtained. We partition the horizontal field evenly into w columns and distribute the vertical field uniformly into h rows. Consequently, post-projection, a depth image bearing a resolution of w × h can be secured, where the horizontal resolution is proportional to the sensor’s rotation rate, and the vertical resolution is proportional to the number of physical layers [44].
An illustration of a specific frame of the depth images under noticeably snowy conditions, along with the corresponding RGB images featuring identical objects, can be found in Figure 2. A close observation reveals that the relevant objects are well reflected and the snow noises, bearing resemblance to ‘salt and pepper’ speckles, are prominently displayed in the depth image. The position of each pixel reflects the frontal-view projection of the points within the point clouds, while the grayscale pixel value signifies the distance between the points and the observing vehicle. Smaller pixel values (darker) imply greater distances, whereas higher pixel values (brighter) suggest closer distances. A discrepancy is worth mentioning that due to the limited vertical FOV and its relatively high mounting position on the car’s roof of the LiDAR used in the sample dataset, the black car in the immediate right vicinity of the ego vehicle partly falls within the “blind spot” and is not clearly delineated in the depth image. Instead, only a few high-intensity signal points from the window glass are visible.
To obtain depth images under clear conditions, we apply the DSOR filter [26] onto snow datasets. In this context, we invert the typical approach of creating synthetic weather conditions in datasets for training, instead of generating an artificial clear dataset. Given that filters cannot guarantee absolute precision and recall in the de-snowing process, the filtered result cannot be considered equivalent to the ground truth. However, the stringent filter still provides a valuable sample for unpaired training. Among the reasons we choose DSOR over alternative filters are its rapid processing speed and its excellent capacity to retain as many environmental elements as possible when the filter parameters are set to an uncompromising level.

3.2. LiDAR Depth Images GAN

In this research, an enhanced iteration of CycleGAN [22] serves as the foundational structure for our proposed model, LiDAR depth images GAN (L-DIG). As illustrated in Figure 3, the comprehensive blueprint of our model is depicted. Due to the features of CycleGAN, both the de-snowing ( s n o w c l e a r ) and the fake snow generation ( c l e a r s n o w ) are completed at the same time for one set of data. Table 1 provides the annotations of all the symbols and alphabet designations used in the model architecture. In our model, A and B symbolize two sets of data flow in the forward and backward cycle, respectively, and Snow A and Clear B are the inputs that we provide in the form of depth images. C subscripts correspond to clear weather conditions, whereas S subscripts are used to indicate snowy conditions. The generators responsible for transitioning between snowy to clear and clear to snowy states are represented by G S C and G C S , respectively. Meanwhile, the discriminators are expressed as D A and D B . The two reconstructions are GAN features to keep the translation from overfitting. We provide the pseudo-code of L-DIG in Algorithm 1.

3.3. Pixel-Attention Discriminators

We designed a new discriminator structure to enable the model to recognize the snow noise points more accurately. The discriminators D A and D B are each composed of two parts: N-layer Discriminators D n A and D n B with 3 convolutional layers, and pixel discriminators D p A and D p B . The N-layer discriminators concentrate on relevant objects within the scene while the pixel discriminators scrutinize each pixel individually to ascertain its authenticity through a 1 × 1 patch [22]. This approach contributes to a minor disturbance to the binary discriminator’s threshold, elevating the criteria to achieve a 1 (approved) rather than a 0 (rejected) to a more stringent level for isolated noise points. This strategy markedly enhances the de-snowing effect, particularly in areas surrounding the ego vehicle, where dispersed snow points are densely packed. The pixel-attention discriminators undergo training for several epochs subsequent to the stabilization of the model training with N-layer discriminators, as shown in Figure 3 and Algorithm 1.
Algorithm 1 LiDAR Depth Images GAN (L-DIG)
Require: Training data pairs ( A , B )    ⊲ Snow A and Clear B
Ensure: Generator networks G S C and G C S , N-Layer Discriminators D n A , D n B , and Pixel Discriminators D p A , D p B
 1:
Initialize generators G S C , G C S and N-Layer Discriminators D n A , D n B
 2:
Define loss functions including GAN, cycle, depth, and SSIM loss
 3:
Define optimizers for generators and discriminators
 4:
while epoch ≤ (total_epochs - continued_epochs) do
 5:
   for each data pair ( A , B ) in data_loader do
 6:
     Generate fake images: F A = G C S ( B ) , F B = G S C ( A )
 7:
     Generate reconstructed images: R e c A = G C S ( F B ) , R e c B = G S C ( F A )
 8:
     Compute GAN, cycle, depth, and SSIM loss
 9:
     Update discriminators D n A , D n B and generators G S C , G C S
10:
   end for
11:
end while
12:
Initialize Pixel Discriminators D p A , D p B
13:
while (total_epochs - continued_epochs) < epoch ≤ total_epochs do
14:
   for each data pair ( A , B ) in data_loader do
15:
     Use the same generators to produce fake images as in previous training
16:
     Compute GAN, cycle, depth, and SSIM loss
17:
     Update discriminators D p A , D p B , D n A , D n B and generators G S C , G C S
18:
   end for
19:
end while

3.4. Loss Function

3.4.1. Depth Loss

Scale ambiguity poses a challenge for depth images, necessitating the use of loss functions resilient to rough estimations [45]. Taking a cue from the Fine network [46], we crafted a depth loss, L d e p t h , that is integrated into the training cycles to uphold consistency in the scale of depth images, as represented in (4).
L d e p t h = 1 n i d i ^ d i 2 λ d e p t h n 2 i d ^ i d i 2
In the given formula, d i ^ and d i symbolize the reconstructed and initial depth, respectively, while the hyperparameter λ d e p t h governs the scale invariance. We assign λ d e p t h = 1 to achieve complete scale invariance. This is due to our objective of preserving the translated point cloud as similar to the original as possible, exclusive of the snow, and safeguarding the relevant objects and environmental elements from distortions in shape and size.

3.4.2. SSIM Loss

As the point cloud translation occurs on the scale of the entire scene, it sometimes involves objects and structures that are partially obscured or incomplete, leading to the model’s suboptimal comprehension of these elements. This can cause distortions or alterations in the original forms, particularly in environmental features. To mitigate this, an SSIM (structural similarity index measure) loss, depicted in (5) and (6), has been incorporated into the training cycle to aid in preserving structural consistency.
S S I M ( N , N ^ ) = ( 2 μ N μ N ^ + c 1 ) ( 2 σ N N ^ + c 2 ) ( μ N 2 + μ N ^ 2 + c 1 ) ( σ N 2 + σ N ^ 2 + c 2 )
L s s i m = 1 S S I M ( N ^ , N )
where N is the normalized image tensor of the original real image, N ^ is the normalized image tensor of the reconstructed image, μ N ^ is the average of N ^ , μ N is the average of N, σ N ^ 2 is the variance of N ^ , σ N 2 is the variance of N, σ N ^ N is the covariance of N ^ and N, c 1 and c 22 are two variables to stabilize the division with a weak denominator.
The SSIM loss computation is performed post a subtraction by 1, due to the fact that SSIM loss gauges similarity while the training mechanism is geared towards attaining minimum values. Consequently, we aim to lower the difference by training 1 minus the SSIM function. This also elucidates the necessity for prior normalization on the image tensor between [0, 1]. Meanwhile, it is essential to maintain a relatively low λ s weight setting on the SSIM loss to prevent the model from becoming overly rigid, thereby obstructing any desired translation.

3.4.3. Cycle Consistency Loss

We employ the cycle consistency loss (7) derived from CycleGAN, aimed at maintaining the created depth images closely aligned with the original domain. Provided a minimal variation in the background during the translation, the weight λ c of cycle consistency loss can be set as equal to the customized depth loss.
L c y c = G S C ( G C S ( S B ) ) S B + G C S ( G S C ( C A ) ) C A

3.4.4. Overall Loss Function

Upon integrating the conventional GAN adversarial losses between the clear and snowy data, denoted by L G A N ( G S C , D A , S , C ) and L G A N ( G C S , D B , C , S ) , we arrive at the comprehensive objective loss:
L ( G S C , G C S , D A , D B ) = L G A N ( G S C , D A , S , C ) + L G A N ( G C S , D B , C , S ) + λ c L c y c + λ d L d e p t h + λ s L s s i m
where λ c , λ d , and λ s denote the weight coefficients of cycle consistency loss, depth loss, and SSIM loss, respectively. The higher the weight, the larger the influence the corresponding loss function has on the model.

3.5. OPTICS 3D Clustering Algorithm

OPTICS exhibits great adaptability to variable densities, presents lower parameter sensitivity, exhibits hierarchical relationships among clusters, and possesses improved outlier detection capabilities. These features of OPTICS largely match with snow point behaviors in point clouds and make it a nice evaluation method on snow datasets. By applying an adaptive clustering principle, we significantly simplify the interpretation and extraction of clusters without calling excessive manual parameter tuning.
We produce the following seven metrics based on the OPTICS algorithm:
  • Noise Number: Points without any neighbor points within a designated range (solitary points) are considered noise points, mostly snowflakes. A decrease in noise number is one of the most direct indicators of an effective de-snowing performance.
  • Cluster Number: A main output of the algorithm, representing groups of data points that are closely related based on their reachability.
  • Reachability Distance: The smallest distance required to connect point A to point B via a path of points that satisfy the density criteria. Normally, the average reachability distance would rise along with larger cluster numbers.
  • Inter-Cluster Distances: The concept here involves identifying the centroid, or the average point, of each cluster, and subsequently computing the distance between every possible pair of centroids. Should there be an increase in the average of these distances, it would suggest a reduction in the number of clusters and a more dispersed cluster distribution. In the context of our study, such a pattern could be interpreted as an effect of de-snowing.
  • Size of Clusters: This is essentially determined by the average number of points each cluster holds. Under conditions dominated by scattered snow, the snow noise points tend to form numerous small-scale clusters. Their elimination, consequently, leads to an increase in the average size of the clusters.
  • Silhouette Score: Measures the cohesion within clusters and the separation between clusters. A silhouette score close to 1 indicates a good clustering quality, while a score close to −1 indicates poor clustering. A lower silhouette score is commonly observed in snowy conditions due to the more overlap between clusters.
  • Davies-Bouldin Index (DBI): Measures the ratio of within-cluster scatter to between-cluster separation and assesses the quality of the overall cluster separation. A lower Davies-Bouldin index indicates better clustering, with zero being the ideal value. Snow conditions with many noise points or swirl clusters exhibit higher values of DBI.
We present three example scenes in Figure 4, from a slight snowfall condition to a normal snowfall condition, as well as a fierce snow swirl condition to illustrate how the OPTICS algorithm evaluates snow conditions, with their metrics summarized in Table 2. From the statistics, we can tell that with the increase in snow level, the noise number and cluster number of the driving scene are gradually rising, along with the average reachability distances as expected. With fewer noise points and fewer snow clusters, the average inter-cluster distances and the average sizes of clusters decrease correspondingly. The tendencies toward deteriorated clustering and increased overlaps, as indicated by the DBI and silhouette score, are also consistent with conditions of heavier snow conditions. This validation proves the OPTICS algorithm’s capability of evaluating the change of snow conditions in a LiDAR point cloud.
In order to more effectively illustrate the differences between mild and heavy snow conditions in terms of reachability distances, inter-cluster distances, and cluster sizes, we provide violin plots for these three metrics in Figure 5, positioning mild snow on the left and heavy snow on the right in each set. These plots not only show their quartiles but more crucially, delineate the differences in distribution. It can be observed that as snow conditions get heavier, the distribution in all three metrics exhibits somewhat abrupt curves with sharper edges and sudden shifts, indicating the inherent disarray characteristic of heavy snow. Generally, scenes with less snow presence display a relatively uniform and smooth distribution, as depicted in the violin plots, with a lower skewness value [47], offering another angle to assess the changes in snow conditions.

4. Experiments and Results

4.1. Experiments

We conducted experiments with the trained models on two different conditions: (1) Mild snow conditions: snowfalls only without snow swirls; (2) Fierce snow conditions: both snowfalls and snow swirls. We first conducted the experiment on snowfall-only conditions to examine the performance of scattered noise point capture. In the meantime, this less occlusion condition provides a better opportunity to check how well the original environmental structures have been maintained, so as to affirm the model’s ability for accurate snow capturing. Then, the same experiment was conducted on conditions with both snowfalls and snow swirls, to comprehensively present the model’s ability to handle highly adverse conditions.
To guarantee the realism of the snow effect, we utilized the well-known dataset specializing in snow conditions within an autonomous driving context—the CADC dataset [2]. This dataset encompasses over 7000 frames of LiDAR point cloud data gathered during the winter season in Waterloo, Ontario. The driving scenarios span both urban and suburban settings, incorporate high and low-speed conditions, and cover a range of snowfall intensities and heavy snow accumulation situations. The LiDAR system implemented in the CADC dataset has a vertical FOV of 40 , with a range from +10 to −30 , and a horizontal FOV of 360 .
Training, testing, and data processing were conducted utilizing the Pytorch framework. We initially examined all possible combinations of ResNet residual blocks (ranging from 4 to 9) in G S C and G C S , and convolutional layers (ranging from 1 to 4) in D n A and D n B , and identified the most optimal combination for our model. When variables are kept constant, a combination of four ResNet residual blocks in G C S , G S C and two downsampling convolutional layers in D n A and D n B produce the most superior translation result.
Square-shaped samples randomly cropped from the depth images are input to two NVIDIA RTX 3090Ti graphics cards with a batch size of eight for training. In the second half of the N-Layer discriminator stage training, we adhere to a linearly declining learning rate schedule until the process converges.
The quantitative analysis is conducted based on 500 samples under mild snowfall conditions and the other 500 samples under fierce snow swirl conditions out of the testing dataset. The reported metrics in the following results all mean the average values.

4.2. Results

4.2.1. Mild Snow Conditions

Figure 6 shows the translation results of our model under mild snow conditions, which means the majority of the snow is scattered noise points without the snow swirl phenomenon. (a), (b), and (c) sets show three typical scenarios with the top row being the original snow scene from CADC, the middle row being the de-snowed results, and the bottom row being the fake snow results. Each of the three scenarios presented features an overall BEV on the left, while the right shows a magnified third-person view of the point cloud’s central region, where the ego vehicle is situated.
As indicated by the red arrows and encircled by red boxes, it is clear that the ‘salt-and-pepper’ noise points have been largely erased, with key environmental features left unaltered. Essential components, like vehicles (outlined in green) and pedestrians (highlighted in orange), are not only well preserved but also exhibit a level of point enhancement, as demonstrated in the de-snow (a) set. Moreover, the road sign enclosed in the red box of (a) which was partially obscured by snow points in the earlier image, seems to be better defined, a testament to the deep scene comprehension facilitated by our model.
The quantitative analyses are presented in Table 3. The noticeable reduction in the average noise number, cluster count, and overall reachability distances in the de-snowed results strongly suggests the effectiveness of the de-snowing process. As the majority of clusters now comprise object points and environmental features that are more densely and uniformly packed, the average inter-cluster distances, and average cluster sizes naturally increase. This shift in cluster characteristics is a byproduct of fewer, but more meaningful, clusters primarily representing substantive elements of the environment rather than scattered snow points. Similarly, the declines in the DBI and silhouette score are in line with our expectations for the de-snowing process.
In the violin plots of Figure 7, the colored data on the left represents de-snowed data, while the gray data on the right serves as a comparison from the CADC dataset. This arrangement is consistent across all subsequent violin plots. A glance at the better evenness within the cluster distribution on the left half of each violin plot reveals the improvement of the de-snowing process compared to the slightly skewed distribution on the right. This observation is further substantiated by the lower skewness of the de-snowed distributions. Our calculations show that for the reachability distances, inter-cluster distances, and sizes of clusters, the skewness values for the de-snow data are 8.11 , 0.23 , and 16.10 , respectively, while for the CADC data, these values are 9.64 , 0.30 , and 21.49 . Note that the median reachability distance of the de-snow is a little bit higher than with snow. This small anomaly originates from a few detached clusters at a remote distance after de-snowing, which can be seen from very few sample points exceeding the upper limit of the y-axis.
For fake snow generations, as seen in the bottom row of Figure 6, the scattered snow features are noticeably reproduced, and there is an apparent enhancement as the number of noise points is higher. This is in line with the noticeable increase in cluster number and DBI, as well as the reduction in cluster sizes, as presented in Table 3. The artificially generated snow demonstrates a remarkable replication capacity, as evidenced by the highly alike violin plots (left and right) in Figure 8, including the quartile lines. The degree of skewness (8.87, 0.33, and 22.43) is remarkably close to the previously mentioned CADC snow skewness (9.64, 0.30, and 21.49), further attesting to the model’s ability to accurately reproduce snow effects.

4.2.2. Fierce Snow Conditions

Figure 9 demonstrates the translation outcomes of our model under intense snow conditions, characterized by the presence of snow swirls around the ego vehicle. Three distinctive scenarios have been chosen for illustration, and are presented in the same format as in Figure 6. In these harsh snowy conditions where the snowfall has dramatically increased, it becomes easier to observe that the vibrantly colored airborne snowdrifts (highlighted in shades of red, green, yellow, and cyan) have been substantially mitigated, as indicated by the red arrows.
Under these severe snow circumstances featuring dense snow swirl clusters, our attention is more on the noise reduction near the ego vehicle, as indicated by the red boxes, instead of entirely eradicating the snow swirls, as this could lead to a loss of important environmental elements. We strive for a balance between significant snow removal and effective preservation of objects like the vehicles shown in the green boxes. Simultaneously, a certain degree of point cloud restoration can also be observed near the central ground rings. This can be credited to the profound comprehension of the scene by the translation model.
Table 4 provides a quantitative representation of the translation model’s performance under extreme snow conditions. Given that the translation effects are applied to more clusters spanning the entire scene during heavy snowfall, all metrics largely veer towards less noise and tidier clustering results in the de-snowing task. From Figure 10, we can tell that the shifts in quartile lines are less prominent, which can be attributed to the fact that snow swirls typically have capacities similar to those of object clusters. Nevertheless, the efficacy of the de-snowing process is evidenced by the smoother and more consolidated distributions in the violin plots. This assertion is additionally validated by the slightly improved skewness of the de-snowed data which stand at 8.87, 0.38, and 28.04, respectively. Conversely, for the CADC data, these values are 10.45, 0.42, and 32.76.
As can be observed from the bottom row of Figure 9, our model effectively replicates airborne snowdrifts and the snow swirls surrounding the ego vehicle. However, the model exhibits a slight restraint at the point cloud’s central region. This outcome results from our strong focus on comprehending the driving scene, which is to avoid the broken integrity obstructed by the extremely dense snow swirl clusters. This situation leads to a somewhat compromised snow imitation, as corroborated by the close statistical outcomes in Table 4, which do not exhibit any significant jumps. Still, the near-symmetrical violin plots in Figure 11 further substantiate the successful emulation of snow effects. The smoother edges of reachability distances and the more concentrated distribution in cluster sizes of the imitation snow hint at a feature of reduced noise at the center. The skewness values associated with the artificially generated heavy snow are 9.95, 0.42, and 31.75, respectively. These values closely mirror those obtained from actual data, which are 10.45, 0.42, and 32.76, respectively. This statistical similarity provides additional evidence, affirming our model’s effectiveness in the task of synthetic snow generation.

4.3. Ablation Study

To affirm the significance of our model’s key components, we conduct an ablation study using the de-snow model under mild snow conditions. This study investigates the impact of the absence of the pixel-attention discriminator, SSIM loss, depth loss, and the basic CycleGAN. Additionally, we examine a training pair with a considerable domain gap. For this purpose, we select 6000 frames from the LIBRE dataset [1], which was collected under clear conditions in Nagoya, Japan’s urban area. This choice serves as a representative due to the substantial domain disparity between Canada and Japan in terms of scenario layouts and traffic patterns. The CADC dataset contains a large portion of suburban scenarios with fewer buildings and more vegetation, which hardly appears in the LIBRE dataset. Table 5 presents the results, using our proposed model as a reference.
The absence of the pixel-attention discriminator results in an immediate degradation in the performance, as evidenced by the increased noise number and reachability distance. Failing to remove a certain amount of solitary noise points substantiates the importance of the pixel-attention discriminator in de-snowing.
More noise points are observed in the scenario without SSIM loss. Apart from the slightly reduced cluster number, other metrics, especially the elevated reachability distance, indicate a breakdown in structural integrity during the translation process. A primary objective of our model is to maintain the crucial objects and environmental elements as effectively as possible, thus affirming the critical role of SSIM loss in our model.
The scenario without depth loss indicates a complete failure in de-snowing, as evidenced by the significant plummeting in all metrics toward noisy and poor clustering. The cause of this failure lies in the unique properties of depth images, which are highly sensitive to non-linear scale changes during the conversion back to point clouds. Consequently, the depth loss forms the cornerstone of our translation model based on depth images.
In the basic CycleGAN model, the mediocre statistics could be interpreted as an utter ineffectiveness in point cloud translation, without managing to preserve the original states either. This result underscores the necessity of all the components in our model for achieving successful translation outcomes.
Finally, when trained on datasets with a substantial domain gap, the model does not yield satisfactory de-snowing performance. This is suggested by the exceedingly high noise number, reachability distances, and low cluster sizes, at least under the same parameter settings as before. The unjustifiably high noise and cluster numbers are the result of poor clustering, which is corroborated by the exceedingly high DBI. This result, derived under extreme conditions, serves to confirm our judicious decision to generate unpaired clear data with filters. However, it does not necessarily suggest that our model lacks generality. Despite this, the model’s robustness against domain gaps does stand as a major limitation of our current translation model.

5. Conclusions

In this research, we introduced a GAN-driven model capable of translating LiDAR point cloud data, encompassing both the removal of snow and artificial snow production. Utilizing depth image priors of point clouds, our model was trained on unpaired datasets and supplemented with depth loss and SSIM loss functions to ensure scale and structure consistencies. Furthermore, we crafted a novel discriminator structure with an emphasis on pixels. This feature was integrated alongside the original convolution layers, thereby enhancing the snow removal capability in the vicinity of the ego vehicle. Experiments carried out using authentic snow conditions from the CADC dataset revealed a profound comprehension of snow characteristics and driving scenes as well as exceptional performance in snow removal and snow reproduction. The 3D clustering results from the OPTICS algorithm and their corresponding violin plots evidently prove the successful translation between snow and clear conditions.
LiDAR point cloud processing under snowy conditions has consistently faced the challenge of lacking reliable snow-affected data. Given the difficulty in acquiring paired or quasi-paired data under both snowy and clear conditions, our current model must strike a balance between the strength of translation and model stability, which subsequently leads to domain sensitivity. Moreover, the limited resources of the CADC dataset intensify the adversity for training and testing. To address these limitations, our future goal is to develop the capability to generate high-quality paired data under snowy conditions. This aim is to augment the LiDAR point cloud with snow based on a deep understanding of the driving scene, with the ultimate intention of preserving the original state of the scene to the greatest extent possible.

Author Contributions

Conceptualization, Y.Z. and H.Y.; methodology, Y.Z. and Y.N.; formal analysis, M.D., K.O. and Y.F.; manuscript preparation, Y.Z.; review and editing, M.D. and K.O.; supervision, K.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Nagoya University and JST, the establishment of university fellowships toward the creation of science technology innovation, Grant Number JPMJFS2120, and JSPS KAKENHI Grant Number JP21H04892 and JP21K12073.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LiDARLight Detection And Ranging
GANGenerative Adversarial Networks
L-DIGLiDAR Depth Images GAN
ADSAutonomous Driving Systems
CADCCanadian Adverse Driving Conditions
OPTICSOrdering Points To Identify the Clustering Structure
DBSCANDensity-Based Spatial Clustering of Applications with Noise
SSIMStructural Similarity Index
DRORDynamic Radius Outlier Removal
DSORDynamic Statistical Outlier Removal
PGMPolar Grid Map
FOVField of View
DBIDavies-Bouldin Index

References

  1. Carballo, A.; Lambert, J.; Monrroy, A.; Wong, D.; Narksri, P.; Kitsukawa, Y.; Takeuchi, E.; Kato, S.; Takeda, K. LIBRE: The multiple 3D LiDAR dataset. In Proceedings of the Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA, 19 October–13 November 2020; pp. 1094–1101. [Google Scholar]
  2. Pitropov, M.; Garcia, D.E.; Rebello, J.; Smart, M.; Wang, C.; Czarnecki, K.; Waslander, S. Canadian adverse driving conditions dataset. Int. J. Robot. Res. 2021, 40, 681–690. [Google Scholar] [CrossRef]
  3. Diaz-Ruiz, C.A.; Xia, Y.; You, Y.; Nino, J.; Chen, J.; Monica, J.; Chen, X.; Luo, K.; Wang, Y.; Emond, M.; et al. Ithaca365: Dataset and Driving Perception Under Repeated and Challenging Weather Conditions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 21383–21392. [Google Scholar]
  4. Zhang, Y.; Carballo, A.; Yang, H.; Takeda, K. Perception and sensing for autonomous vehicles under adverse weather conditions: A survey. ISPRS J. Photogramm. Remote Sens. 2023, 196, 146–177. [Google Scholar] [CrossRef]
  5. Guo, A.; Feng, Y.; Chen, Z. LiRTest: Augmenting LiDAR point clouds for automated testing of autonomous driving systems. In Proceedings of the 31st ACM SIGSOFT International Symposium on Software Testing and Analysis, Virtual, 18–22 July 2022; pp. 480–492. [Google Scholar]
  6. Goodin, C.; Carruth, D.; Doude, M.; Hudson, C. Predicting the Influence of Rain on LIDAR in ADAS. Electronics 2019, 8, 89. [Google Scholar] [CrossRef]
  7. Jokela, M.; Kutila, M.; Pyykönen, P. Testing and validation of automotive point-cloud sensors in adverse weather conditions. Appl. Sci. 2019, 9, 2341. [Google Scholar] [CrossRef]
  8. Charron, N.; Phillips, S.; Waslander, S.L. De-noising of lidar point clouds corrupted by snowfall. In Proceedings of the 2018 15th Conference on Computer and Robot Vision (CRV), Toronto, ON, Canada, 8–10 May 2018; pp. 254–261. [Google Scholar]
  9. Le, M.H.; Cheng, C.H.; Liu, D.G. An Efficient Adaptive Noise Removal Filter on Range Images for LiDAR Point Clouds. Electronics 2023, 12, 2150. [Google Scholar] [CrossRef]
  10. Sakaridis, C.; Dai, D.; Van Gool, L. Semantic foggy scene understanding with synthetic data. Int. J. Comput. Vis. 2018, 126, 973–992. [Google Scholar] [CrossRef]
  11. Zhang, H.; Patel, V.M. Densely connected pyramid dehazing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3194–3203. [Google Scholar]
  12. Uřičář, M.; Křížek, P.; Sistu, G.; Yogamani, S. Soilingnet: Soiling detection on automotive surround-view cameras. In Proceedings of the Intelligent Transportation Systems Conference (ITSC), Auckland, New Zealand, 27–30 October 2019; pp. 67–72. [Google Scholar]
  13. Yue, Z.; Xie, J.; Zhao, Q.; Meng, D. Semi-Supervised Video Deraining With Dynamical Rain Generator. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 642–652. [Google Scholar]
  14. Ye, Y.; Chang, Y.; Zhou, H.; Yan, L. Closing the Loop: Joint Rain Generation and Removal via Disentangled Image Translation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 2053–2062. [Google Scholar]
  15. Ancuti, C.; Ancuti, C.O.; Timofte, R. Ntire 2018 challenge on image dehazing: Methods and results. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 891–901. [Google Scholar]
  16. Engin, D.; Genç, A.; Kemal Ekenel, H. Cycle-dehaze: Enhanced cyclegan for single image dehazing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 825–833. [Google Scholar]
  17. Yang, H.; Carballo, A.; Zhang, Y.; Takeda, K. Framework for generation and removal of multiple types of adverse weather from driving scene images. Sensors 2023, 23, 1548. [Google Scholar] [CrossRef] [PubMed]
  18. Uřičář, M.; Sistu, G.; Rashed, H.; Vobecky, A.; Kumar, V.R.; Krizek, P.; Burger, F.; Yogamani, S. Let’s Get Dirty: GAN Based Data Augmentation for Camera Lens Soiling Detection in Autonomous Driving. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Virtual, 5–9 January 2021; pp. 766–775. [Google Scholar]
  19. Japan Automotive Research Institute (JARI). Special Environment Proving Ground.
  20. Laboratoire régional des ponts et chaussées. Site de Clermont-Ferrand.
  21. Heinzler, R.; Piewak, F.; Schindler, P.; Stork, W. Cnn-based lidar point cloud de-noising in adverse weather. IEEE Robot. Autom. Lett. 2020, 5, 2514–2521. [Google Scholar] [CrossRef]
  22. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  23. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  24. Yang, H.; Carballo, A.; Takeda, K. Disentangled Bad Weather Removal GAN for Pedestrian Detection. In Proceedings of the 2022 IEEE 95th Vehicular Technology Conference: (VTC2022-Spring), Helsinki, Finland, 19–22 June 2022; pp. 1–6. [Google Scholar]
  25. Caccia, L.; Van Hoof, H.; Courville, A.; Pineau, J. Deep generative modeling of lidar data. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), The Venetian Macao, Macau, 3–8 November 2019; pp. 5034–5040. [Google Scholar]
  26. Kurup, A.; Bos, J. Dsor: A scalable statistical filter for removing falling snow from lidar point clouds in severe winter weather. arXiv 2021, arXiv:2109.07078. [Google Scholar]
  27. Bergius, J. LiDAR Point Cloud De-Noising for Adverse Weather. Ph.D. Thesis, Halmstad University, Halmstad, Sweden, 2022. [Google Scholar]
  28. Wang, C.; Ji, M.; Wang, J.; Wen, W.; Li, T.; Sun, Y. An improved DBSCAN method for LiDAR data segmentation with automatic Eps estimation. Sensors 2019, 19, 172. [Google Scholar] [CrossRef] [PubMed]
  29. Ankerst, M.; Breunig, M.M.; Kriegel, H.P.; Sander, J. OPTICS: Ordering points to identify the clustering structure. ACM Sigmod Rec. 1999, 28, 49–60. [Google Scholar] [CrossRef]
  30. Schubert, E.; Gertz, M. Improving the Cluster Structure Extracted from OPTICS Plots. In Proceedings of the LWDA, Mannheim, Germany, 22–24 August 2018; pp. 318–329. [Google Scholar]
  31. El Yabroudi, M.; Awedat, K.; Chabaan, R.C.; Abudayyeh, O.; Abdel-Qader, I. Adaptive DBSCAN LiDAR Point Cloud Clustering For Autonomous Driving Applications. In Proceedings of the 2022 IEEE International Conference on Electro Information Technology (eIT), Mankato, MN, USA, 19–21 May 2022; pp. 221–224. [Google Scholar]
  32. Park, J.I.; Park, J.; Kim, K.S. Fast and accurate desnowing algorithm for LiDAR point clouds. IEEE Access 2020, 8, 160202–160212. [Google Scholar] [CrossRef]
  33. Wang, W.; You, X.; Chen, L.; Tian, J.; Tang, F.; Zhang, L. A scalable and accurate de-snowing algorithm for LiDAR point clouds in winter. Remote Sens. 2022, 14, 1468. [Google Scholar] [CrossRef]
  34. Bijelic, M.; Gruber, T.; Ritter, W. A benchmark for lidar sensors in fog: Is detection breaking down? In Proceedings of the Intelligent Vehicles Symposium (IV), Changshu, China, 26–30 June 2018; pp. 760–767. [Google Scholar]
  35. Shamsudin, A.U.; Ohno, K.; Westfechtel, T.; Takahiro, S.; Okada, Y.; Tadokoro, S. Fog removal using laser beam penetration, laser intensity, and geometrical features for 3D measurements in fog-filled room. Adv. Robot. 2016, 30, 729–743. [Google Scholar] [CrossRef]
  36. Hahner, M.; Sakaridis, C.; Bijelic, M.; Heide, F.; Yu, F.; Dai, D.; Van Gool, L. Lidar snowfall simulation for robust 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16364–16374. [Google Scholar]
  37. Kim, T.; Cha, M.; Kim, H.; Lee, J.K.; Kim, J. Learning to discover cross-domain relations with generative adversarial networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 1857–1865. [Google Scholar]
  38. Yi, Z.; Zhang, H.; Tan, P.; Gong, M. Dualgan: Unsupervised dual learning for image-to-image translation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2849–2857. [Google Scholar]
  39. Jaw, D.W.; Huang, S.C.; Kuo, S.Y. DesnowGAN: An efficient single image snow removal framework using cross-resolution lateral connection and GANs. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 1342–1350. [Google Scholar] [CrossRef]
  40. Sallab, A.E.; Sobh, I.; Zahran, M.; Essam, N. LiDAR Sensor modeling and Data augmentation with GANs for Autonomous driving. arXiv 2019, arXiv:1905.07290. [Google Scholar]
  41. Sobh, I.; Amin, L.; Abdelkarim, S.; Elmadawy, K.; Saeed, M.; Abdeltawab, O.; Gamal, M.; El Sallab, A. End-to-End Multi-Modal Sensors Fusion System for Urban Automated Driving. 2018. Available online: https://openreview.net/forum?id=Byx4Xkqjcm (accessed on 15 October 2023).
  42. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? the kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
  43. Lee, J.; Shiotsuka, D.; Nishimori, T.; Nakao, K.; Kamijo, S. GAN-Based LiDAR Translation between Sunny and Adverse Weather for Autonomous Driving and Driving Simulation. Sensors 2022, 22, 5287. [Google Scholar] [CrossRef] [PubMed]
  44. Zyrianov, V.; Zhu, X.; Wang, S. Learning to generate realistic lidar point clouds. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 17–35. [Google Scholar]
  45. Mertan, A.; Duff, D.J.; Unal, G. Single image depth estimation: An overview. Digit. Signal Process. 2022, 123, 103441. [Google Scholar] [CrossRef]
  46. Eigen, D.; Puhrsch, C.; Fergus, R. Depth map prediction from a single image using a multi-scale deep network. Adv. Neural Inf. Process. Syst. 2014, 27. [Google Scholar]
  47. Kokoska, S.; Zwillinger, D. CRC Standard Probability and Statistics Tables and Formulae; CRC Press: Boca Raton, FL, USA, 2000. [Google Scholar]
Figure 1. A frame of the point cloud featuring both dispersed noise points and snow clusters. (a) The original point cloud from the Canadian Adverse Driving Conditions (CADC) dataset [2], with colors representing height—the lighter the color, the greater the distance from the ground. Red boxes annotate scattered snow and snow swirl points. (b) The clustering result of the same point cloud based on the OPTICS (Ordering points to identify the clustering structure) algorithm, where varying colors signify different cluster groups. Objects and structures with regularized forms are aptly segmented into large-scale clusters, while scattered snow noise points constitute smaller clusters with minimal points. The prominent purple cluster at the center represents the snow swirl.
Figure 1. A frame of the point cloud featuring both dispersed noise points and snow clusters. (a) The original point cloud from the Canadian Adverse Driving Conditions (CADC) dataset [2], with colors representing height—the lighter the color, the greater the distance from the ground. Red boxes annotate scattered snow and snow swirl points. (b) The clustering result of the same point cloud based on the OPTICS (Ordering points to identify the clustering structure) algorithm, where varying colors signify different cluster groups. Objects and structures with regularized forms are aptly segmented into large-scale clusters, while scattered snow noise points constitute smaller clusters with minimal points. The prominent purple cluster at the center represents the snow swirl.
Sensors 23 08660 g001
Figure 2. An illustration of a specific frame of the depth image under heavy snow conditions. The middle row displays the depth image, while the top and bottom rows depict corresponding RGB images derived from the CADC(Canadian Adverse Driving Conditions) dataset [2]. These images are captured from multiple cameras targeting different directions around the ego vehicle. Color-coded boxes indicate objects that are present in both the RGB and depth images. Green—A pole with a road sign and traffic lights. Yellow—Three trees. Red—A vehicle at the intersection. Blue—Two pick-up trucks waiting in line. Please note that the images provided do not represent the original resolutions; they have been adjusted to enhance their illustrative capacity.
Figure 2. An illustration of a specific frame of the depth image under heavy snow conditions. The middle row displays the depth image, while the top and bottom rows depict corresponding RGB images derived from the CADC(Canadian Adverse Driving Conditions) dataset [2]. These images are captured from multiple cameras targeting different directions around the ego vehicle. Color-coded boxes indicate objects that are present in both the RGB and depth images. Green—A pole with a road sign and traffic lights. Yellow—Three trees. Red—A vehicle at the intersection. Blue—Two pick-up trucks waiting in line. Please note that the images provided do not represent the original resolutions; they have been adjusted to enhance their illustrative capacity.
Sensors 23 08660 g002
Figure 3. Proposed LiDAR translation model architecture. Datasets A and B are in the form of depth images. G S C and G C S are the generators. D n A and D n B are the N-layer discriminators, and D p A and D p B are the pixel-attention discriminators.
Figure 3. Proposed LiDAR translation model architecture. Datasets A and B are in the form of depth images. G S C and G C S are the generators. D n A and D n B are the N-layer discriminators, and D p A and D p B are the pixel-attention discriminators.
Sensors 23 08660 g003
Figure 4. The clustering results of three example scenes from the OPTICS algorithm, where varying colors signify different cluster groups. (a) Slight snowfall condition with bits and pieces of snow points. (b) Normal snowfall condition with both scattered snow points and snow clusters around the center. (c) Fierce snow swirl condition with huge snow swirl clusters surrounding the ego vehicle.
Figure 4. The clustering results of three example scenes from the OPTICS algorithm, where varying colors signify different cluster groups. (a) Slight snowfall condition with bits and pieces of snow points. (b) Normal snowfall condition with both scattered snow points and snow clusters around the center. (c) Fierce snow swirl condition with huge snow swirl clusters surrounding the ego vehicle.
Sensors 23 08660 g004
Figure 5. Violin plots for the comparison between mild snow and heavy snow. Limits on the y-axes are set for a better illustration of the distributions.
Figure 5. Violin plots for the comparison between mild snow and heavy snow. Limits on the y-axes are set for a better illustration of the distributions.
Sensors 23 08660 g005
Figure 6. Qualitative results of point cloud translations in scattered snow conditions (without snow swirls). Colors are encoded by height. The top row is the snow conditions from the CADC dataset, the middle row is our de-snowed results, and the bottom row is the fake snow results obtained from our model. Three scenarios are presented, with the left sides of each set being the overall BEV scenes and the right sides of each set being the enlarged third-person view center part around the ego vehicle. Red boxes and arrows denote the locations where snow’s effects are alleviated. Green boxes annotate vehicles. Orange boxes annotate pedestrians.
Figure 6. Qualitative results of point cloud translations in scattered snow conditions (without snow swirls). Colors are encoded by height. The top row is the snow conditions from the CADC dataset, the middle row is our de-snowed results, and the bottom row is the fake snow results obtained from our model. Three scenarios are presented, with the left sides of each set being the overall BEV scenes and the right sides of each set being the enlarged third-person view center part around the ego vehicle. Red boxes and arrows denote the locations where snow’s effects are alleviated. Green boxes annotate vehicles. Orange boxes annotate pedestrians.
Sensors 23 08660 g006
Figure 7. Violin plots for de-snow results under mild snow conditions. Limits on the y-axes are set for a better illustration of the distributions.
Figure 7. Violin plots for de-snow results under mild snow conditions. Limits on the y-axes are set for a better illustration of the distributions.
Sensors 23 08660 g007
Figure 8. Violin plots for fake snow results under mild snow conditions. Limits on the y-axes are set for a better illustration of the distributions.
Figure 8. Violin plots for fake snow results under mild snow conditions. Limits on the y-axes are set for a better illustration of the distributions.
Sensors 23 08660 g008
Figure 9. Qualitative results of point cloud translations in fierce snow conditions (with snow swirls). Colors are encoded by height. The top row is the snow conditions from the CADC dataset, the medium row is our de-snowed results, and the bottom row is the fake snow results obtained from our model. Three scenarios are presented, with the left sides of each set being the overall BEV scenes and the right sides of each set being the enlarged third-person view center part around the ego vehicle. Red boxes and arrows denote the locations where snow’s effects are alleviated. Green boxes annotate vehicles.
Figure 9. Qualitative results of point cloud translations in fierce snow conditions (with snow swirls). Colors are encoded by height. The top row is the snow conditions from the CADC dataset, the medium row is our de-snowed results, and the bottom row is the fake snow results obtained from our model. Three scenarios are presented, with the left sides of each set being the overall BEV scenes and the right sides of each set being the enlarged third-person view center part around the ego vehicle. Red boxes and arrows denote the locations where snow’s effects are alleviated. Green boxes annotate vehicles.
Sensors 23 08660 g009
Figure 10. Violin plots for de-snow results under fierce snow conditions with swirls. Limits on the y-axes are set for a better illustration of the distributions.
Figure 10. Violin plots for de-snow results under fierce snow conditions with swirls. Limits on the y-axes are set for a better illustration of the distributions.
Sensors 23 08660 g010
Figure 11. Violin plots for fake snow results under fierce snow conditions with swirls. Limits on the y-axes are set for a better illustration of the distributions.
Figure 11. Violin plots for fake snow results under fierce snow conditions with swirls. Limits on the y-axes are set for a better illustration of the distributions.
Sensors 23 08660 g011
Table 1. Annotation of symbols in the model architecture.
Table 1. Annotation of symbols in the model architecture.
SymbolsAnnotationsSymbolsAnnotations
Adata flow in the forward cycleBdata flow in the backward cycle
Ssnow conditionsCclear conditions
Snow Ainput dataset with snow conditionsClear Binput dataset under clear conditions
Snow Bmodel generated snow resultsClear Amodel generated clear results
G S C generator for snow to clear G C S generator for clear to snow
D A discriminator for judging generated clear results D B discriminator for judging generated snow results
D n A N-layer Discriminator module in D A D n B N-layer Discriminator module in D B
D p A pixel-attention discriminator module in D A D p B pixel-attention discriminator module in D B
R e c A reconstructed snow results R e c B reconstructed clear results
Table 2. The 3D clustering metrics from the OPTICS algorithm for ascending snow levels.
Table 2. The 3D clustering metrics from the OPTICS algorithm for ascending snow levels.
Noise NumberCluster NumberReachability Distances Avg.Size of Clusters Avg.Inter-Cluster Distances Avg.Davies-Bouldin IndexSilhouette Score
Slight snowfall18617610.333654.027917.63734.0110−0.1653
Normal snowfall23199820.373752.882014.25975.8451−0.2246
Fierce snow swirl294211650.416646.109012.34948.4671−0.2772
Table 3. The 3D clustering metrics (avg.) from the OPTICS algorithm under scattered snow conditions.
Table 3. The 3D clustering metrics (avg.) from the OPTICS algorithm under scattered snow conditions.
Noise NumberCluster NumberReachability Distances Inter-Cluster DistancesSize of ClustersDavies-Bouldin IndexSilhouette Score
CADC2865.07964.230.407651.426214.16034.0279−0.3011
De-snow2689.46792.230.376653.916315.83663.7427−0.2111
Fake snow3025.061011.040.433951.144712.90284.7175−0.2730
Table 4. The 3D clustering metrics (avg.) from the OPTICS algorithm under snow swirls conditions.
Table 4. The 3D clustering metrics (avg.) from the OPTICS algorithm under snow swirls conditions.
Noise NumberCluster NumberReachability Distances Inter-Cluster DistancesSize of ClustersDavies-Bouldin IndexSilhouette Score
CADC2598.17954.240.380545.924514.97505.4138−0.3688
De-snow1606.23639.990.329347.439224.21103.9073−0.3109
Fake snow3124998.770.435844.304413.71646.0746−0.3661
Table 5. Ablation study based on the de-snow model under mild snow conditions.
Table 5. Ablation study based on the de-snow model under mild snow conditions.
Noise NumberCluster NumberReachability Distances Inter-Cluster DistancesSize of ClustersDavies-Bouldin IndexSilhouette Score
Proposed model2689.46792.230.376653.916315.83663.7427−0.2111
w/o p i x e l D 2882.94761.250.437352.485016.30894.0880−0.3532
w/o ssim loss2935.19772.250.451153.170015.77944.1745−0.3394
w/o depth loss3879.131097.30.559746.18199.99855.4330−0.2990
Basic CycleGAN2891.98833.470.486349.977414.04905.1194−0.3638
Domain gap5031.851132.720.766347.212410.18967.1288−0.2825
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Ding, M.; Yang, H.; Niu, Y.; Feng, Y.; Ohtani, K.; Takeda, K. L-DIG: A GAN-Based Method for LiDAR Point Cloud Processing under Snow Driving Conditions. Sensors 2023, 23, 8660. https://doi.org/10.3390/s23218660

AMA Style

Zhang Y, Ding M, Yang H, Niu Y, Feng Y, Ohtani K, Takeda K. L-DIG: A GAN-Based Method for LiDAR Point Cloud Processing under Snow Driving Conditions. Sensors. 2023; 23(21):8660. https://doi.org/10.3390/s23218660

Chicago/Turabian Style

Zhang, Yuxiao, Ming Ding, Hanting Yang, Yingjie Niu, Yan Feng, Kento Ohtani, and Kazuya Takeda. 2023. "L-DIG: A GAN-Based Method for LiDAR Point Cloud Processing under Snow Driving Conditions" Sensors 23, no. 21: 8660. https://doi.org/10.3390/s23218660

APA Style

Zhang, Y., Ding, M., Yang, H., Niu, Y., Feng, Y., Ohtani, K., & Takeda, K. (2023). L-DIG: A GAN-Based Method for LiDAR Point Cloud Processing under Snow Driving Conditions. Sensors, 23(21), 8660. https://doi.org/10.3390/s23218660

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop