Next Article in Journal
Layout Comparison and Parameter Optimization of Supercritical Carbon Dioxide Coal-Fired Power Generation Systems under Environmental and Economic Objectives
Previous Article in Journal
Research on Risk Contagion among Financial Submarkets in China Based on Complex Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Transfer Network with Multi-Space Dynamic Distribution Adaptation for Bearing Fault Diagnosis

1
School of Electronic Information, Hangzhou Dianzi University, Hangzhou 310018, China
2
Zhejiang Provincial Key Lab of Equipment Electronics, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(8), 1122; https://doi.org/10.3390/e24081122
Submission received: 4 July 2022 / Revised: 4 August 2022 / Accepted: 13 August 2022 / Published: 15 August 2022

Abstract

:
Domain adaptation-based bearing fault diagnosis methods have recently received high attention. However, the extracted features in these methods fail to adequately represent fault information due to the versatility of the work scenario. Moreover, most existing adaptive methods attempt to align the feature space of domains by calculating the sum of marginal distribution distance and conditional distribution distance, without considering variable cross-domain diagnostic scenarios that provide significant cues for fault diagnosis. To address the above problems, we propose a deep convolutional multi-space dynamic distribution adaptation (DCMSDA) model, which consists of two core components: two feature extraction modules and a dynamic distribution adaptation module. Technically, a multi-space structure is proposed in the feature extraction module to fully extract fault features of the marginal distribution and conditional distribution. In addition, the dynamic distribution adaptation module utilizes different metrics to capture distribution discrepancies, as well as an adaptive coefficient to dynamically measure the alignment proportion in complex cross-domain scenarios. This study compares our method with other advanced methods, in detail. The experimental results show that the proposed method has excellent diagnosis performance and generalization performance. Furthermore, the results further demonstrate the effectiveness of each transfer module proposed in our model.

1. Introduction

Bearing is an essential component of rotating machinery and plays a key role in industry. Due to the complex working environment and endless work, bearing failures cannot be avoided. Therefore, in order to ensure normal industrial production, it is important to accurately diagnose bearing faults. Deep neural network based methods have shown excellent performance in fault diagnosis tasks for rotating machinery [1,2,3] and other fields [4] due to their powerful feature capture capabilities. According to [5], the networks used in bearing fault diagnosis can be roughly divided into convolutional neural network (CNN) [6], auto encoder (AE) [7] and deep belief network (DBN) [8]. Although the positive results have been achieved, they still suffer from a problem in real-world scenarios, i.e., the domain shift. Specifically, the distribution of bearing training data and test data collected from variable operating conditions differs significantly. As a result, when a model trained on the training data is deployed to test data from different working conditions, the performance of the deep learning model is greatly limited. In other words, the feature space inconsistency between bearing training data and test data causes the trained model to fail to generalize to the test set.
To counter this domain shift issue, bearing fault diagnosis based on domain adaptation (DA) [5] offers a workable solution, which learns the common fault features of the source and target domain data in different working scenarios. Since the source and target domains have the same fault type in this method, it is possible to learn the classification boundaries of the target domain feature space based on the source domain and the fault classifier, which can successfully separate these domain-invariant features according to category boundaries. Due to the inherent similarity of fault features under different working conditions, the shared features present in both domains allow for some feasibility of domain adaptation [9]. Othman et al. [10] proposed a deep transfer learning model, in which a three-layer autoencoder was used to extract common fault features of the source and target domains for bearing fault identification, with a maximum mean discrepancy (MMD) metric as the main optimization objective. By optimizing the module, the feature space was aligned and the fault classifier utilized the information contained in the domain-invariant features to successfully implement fault diagnosis. An et al. [11] proposed a one-dimensional convolutional neural network to guide common fault representations. To improve network generalizability, the correlated alignment loss (CORAL) was considered as a regularization item to solve the domain shift problem due to inconsistent working conditions. Additionally, to simulate real production conditions, they added Gaussian noise to the input vibration signal. Even when disturbed by the noise, their model still showed superior fault-discriminative performance. Zhu et al. [12] added two additional layers to the CNN to extract task-specific features. By combining multiple Gaussian kernels, the training loss was calculated more accurately. In practice, they deployed different optimization objectives to train the model, also improving accuracy in bearing fault diagnosis. Wan et al. [13] proposed an improved deep network as the feature extractor to remove industrial noise from bearing signals and extract the a priori knowledge. Moreover, they constructed corresponding domain discriminators according to each type of fault. By adversarial training of the domain discriminator and the feature extractor, a more adequate alignment of the feature space was achieved.
Alignment of the feature space is the key to these domain adaptive approaches. As a sub-topic of domain adaptation, joint distribution adaptation (JDA) has been proposed for more levels of feature space alignment. It considers both global feature space alignment and fault category-specific feature space alignment to adapt the fault classifier to changing working conditions. Benefiting from this design, JDA methods perform with high accuracy [13]. However, there are still some problems that need to be addressed, as follows.
  • Most feature representations in JDA methods fail to adequately contain information about the fault types. They often ignore some fine-grained information, such as the size and location of the fault, etc. Although some methods take into account fine-grained information, most usually face problems of slow and difficult convergence [14,15], because the adversarial training process is complicated with two stages;
  • Since the joint distribution cannot be directly measured, most JDA methods calculate the sum of the marginal distribution distance and the conditional distribution distance, which is approximate to the joint distribution distance. Due to the approximation, accurate feature representation is critical in cross-domain scenarios. On one hand, when the bearing signals are collected from widely varying working conditions, the overall feature representations that represent marginal distribution are more important. On the other hand, when from similar conditions, feature representations of conditional distribution specific to each fault class are more critical. However, existing JDA methods have rarely investigated feature representation under various working conditions.
To address these issues, a deep convolutional multi-space dynamic distribution adaptation (DCMSDA) model is proposed for bearing fault diagnosis. To adequately align the sample distribution, we propose a detailed feature extraction scheme and two optimization metrics. Figure 1b shows that marginal distribution adaptation aligns the global feature representations between domains, while Figure 1c shows that conditional distribution adaptation focuses on the category-specific feature representations. Additionally, Figure 1d shows that our model adequately aligns feature representations by automatically selecting the appropriate adaptation method, when faced with an unknown target domain from various working conditions. It can be clearly seen that the inadequate feature alignment problems in JDA methods can be effectively solved, which further enables the domain shift problem to be alleviated. The main contributions of this paper are as follows:
  • A marginal feature extraction module and a multi-space conditional feature extraction module are proposed to guide powerful feature representation of fault information. Based on these modules, the mapping effect of features at multiple scales is achieved;
  • A multi-kernel maximum mean discrepancy (MK-MMD) and a local maximum mean discrepancy (LMMD) are introduced as metrics to adjust the marginal distribution and the conditional distribution, respectively. By optimizing the objectives together, the distribution discrepancies within the extracted features can be reduced;
  • An adaptive coefficient is designed to dynamically measure the alignment proportion of feature representations. It reweights the fault feature representations by the construction of two domain discriminators, improving the generalization performance in complex cross-domain scenarios.
The details of the proposed model and its optimization objective are introduced in Section 2. The comparative experiments and ablation experiments are presented in Section 3. Finally, the conclusions and future work are provided in Section 4.

2. Research Methods

2.1. Proposed Framework

The structure of our DCMSDA model is shown in Figure 2. The deep convolutional multi-space dynamic distribution adaptation network consists of four parts, including the marginal feature extraction module, the conditional feature extraction module, the dynamic distribution adaptation module and the fault classification module. The two feature extraction modules are composed of multiple stacked convolutional layers; the fault classification module is a fully connected layer with Softmax activation function; and the dynamic distribution adaptation module contains two domain discriminators, which consist of three fully connected layers with Sigmoid activation function. Given the raw time–domain signal as input, the marginal and conditional feature extraction modules are used to extract the features of the marginal and conditional distributions, respectively. After that, the fault classification module classifies the extracted fault features and outputs the classification vectors. During the training process, the proposed dynamic distribution adaptation module uses two metrics to measure the distribution discrepancies of domains. Moreover, to calculate the adaptive coefficient, which dynamically guides the feature extraction modules to extract more domain-invariant features, two domain discriminators are constructed to determine the domain to which the above features belong. The primary transfer modules introduced in the model include the feature extraction modules and the dynamic distribution adaptation module, which are described in this Section 2.1. The training process of the model is described in detail in Section 2.2.

2.1.1. Feature Extraction Module

Deep neural network layers make the extracted features more task-specific dependent, while shallow layers learn general features [16,17]. To fully extract feature information, the marginal feature extraction module and the conditional feature extraction module were designed. The features extracted by the marginal feature extraction module were further fed into the conditional feature extraction module. In detail, the marginal feature extraction module is a shallow convolutional neural network, so the extracted features contain more general information with apparent marginal distribution discrepancies, which is called marginal features in this paper. The conditional feature extraction module was built by a multi-space convolutional neural network, which fuses features at different scales to produce conditional features with rich semantic and detailed information. Therefore, conditional features have significant differences to marginal features, enhancing the feature representation in fault classification tasks.
The marginal feature extraction module adopts ResNet-18 1D [5] to extract fault features. The conditional feature extraction module consists of three conditional feature extractors on three spaces, as shown in Figure 3. Here, multiple small convolutional layers with different levels are stacked. To present the model structure more clearly, Table 1 shows details of the implemented conditional feature extraction module. Since the three conditional feature extractors have different depths, convolutional kernel sizes and parameters, richer feature representations with semantic and detailed information can be extracted for the fault diagnosis task. The conditional features learned in the three spaces are fused together to form high-level features, providing informative cues for the fault classifier.

2.1.2. Dynamic Distribution Adaptation Module

The proposed dynamic distribution adaptation module has three interesting elements: MK-MMD focuses on the global distribution discrepancy [18,19] and is suitable for the alignment of the marginal feature distribution; while LMMD focuses on the relationship between two sub-domains within the same category [14], and is appropriate for the conditional feature distribution. In addition, an adaptative coefficient is deployed to intuitively reweight the marginal distribution alignment and the conditional distribution alignment.
MK-MMD is used to optimize the marginal feature distribution, which can be formulated as follows:
L m = d k 2 ( P ( x s ) , P ( x t ) ) E P ( x s ) ( ϕ ( G m f ( x i s ) ) ) E P ( x t ) ( ϕ ( G m f ( x i t ) ) ) k 2
where x i s and x i t represent the i -th source-domain sample and the i -th target-domain sample, respectively, which obey the marginal probability distribution P ( x s ) and P ( x t ) . G m f ( · ) represents the feature representation extracted by marginal feature extractor, k represents reproducing kernel Hilbert space (RKHS), ϕ ( · ) represents the feature mapping of original samples to RKHS, and E represents the mathematical expectation of the two datasets in RKHS.
LMMD is applied to measure the conditional distribution distance on three spaces, which is expressed as follows:
L c = d k 2 ( Q ( x s ) , Q ( x t ) ) 1 C c = 1 C x i s D s ( ω i s , c ϕ ( G c f ( x i s ) ) ) x i t D t ( ω i t , c ϕ ( G c f ( x i t ) ) ) k 2
where Q ( x s ) and Q ( x t ) denote conditional probability distribution, G c f ( · ) represents the feature representation extracted by conditional feature extractor, ω i s , c and ω i t , c represent the weight of the c -th category data x i s and x i t , respectively. The source domain data use the real labels to calculate the corresponding weights, and the target domain data use the pseudo-labels output by the fault classifier to calculate weights.
The two domain binary discriminators classify the domain to which marginal features and conditional features belong. Here, we refer to them as the marginal domain discriminator and the conditional domain discriminator. The optimization objective of the the marginal domain discriminator is calculated by:
L m d = ( G m d ( G m f ( x i ) ) , d i ) = d i log ( G m d ( G m f ( x i ) ) ) ( 1 d i ) log ( 1 G m d ( G m f ( x i ) ) )
where x i is the i -th sample of vibration signals, G m d ( · ) denotes the output of the marginal domain discriminator, and d i is the domain label. Notably, the source domain label and the target domain label are defined as 1 and 0, respectively.
The fused high-level conditional features from three spaces are linearly mapped to the output of the fault classifier. The linear mapping is used as an input of the conditional domain discriminator, which follows some interpretability for the conditional features containing potential category information. The loss function of the conditional domain discriminator can be formulated by:
L c d = ( G c d ( T ( f i , y ^ i ) ) , d i ) = d i log ( G c d ( T ( f i , y ^ i ) ) ) ( 1 d i ) log ( 1 G c d ( T ( f i , y ^ i ) ) )
where f i is the high-level conditional features of the i -th sample, y ^ i is the output probability value of the Softmax function, T ( · , · ) denotes the linear mapping function, and G c d ( · ) denotes the output result of the conditional domain discriminator.
The adaptive coefficient qualitatively and quantitatively combines the classification loss of the two domain discriminators above, which is calculated as follows:
μ = L c d L c d + L m d
The smaller the conditional domain discriminator loss, L c d , the more accurately the domain discriminator classifies the source or target domain. This dynamically adjusts the loss L c according to the conditional distribution of the source domain and target domain. When the conditional distribution varies considerably, the weight for loss will be higher.
Table 2 shows the details of the marginal domain discriminator and the conditional domain discriminator. Layers F1, F2 and F3 are marginal domain discriminator layers. The conditional domain discriminator consists of layers F4, F5 and F6.

2.2. Training Process

The model aims to identify the fault type of the target domain sample by dynamically closing the marginal distribution and conditional distribution in two domains. The training process of the proposed DCMSDA model is detailed below.
Step 1: The labelled source and unlabeled target domain samples are treated as inputs, which are fed into the shared marginal feature extractor. MK-MMD loss for the marginal feature is calculated according to Equation (1).
Step 2: The extracted marginal features are further fed into the conditional feature extractors with three spaces, and LMMD loss for each space is calculated according to Equation (2).
Step 3: With the marginal features and the fused conditional features, the marginal and conditional domain discriminators, output the corresponding classification results, respectively. Then the binary loss for the two discriminators and the adaptive coefficient μ is calculated according to Equations (3)–(5).
Step 4: The fused high-level features are fed into the fault classifier and the source domain classification loss is expressed by:
L y = 1 n i = 1 n j = 1 C 1 y i s = j log e θ j T x i s l = 1 C e θ l T x i s
where n denotes the number of samples in the source domain, C denotes the number of categories, log ( · ) denotes the output of Softmax function, 1 [ · ] denotes the indicator function, and θ 1 , θ 2 , , θ k are learnable parameters in the fault classifier.
The proposed DCMSDA model has four optimization objectives: minimize MK-MMD loss, LMMD loss in three spaces, fault classification loss, and domain classification loss. Combined with these four loss functions, the overall optimization objective is as follows:
L = L y + λ [ μ L m + ( 1 μ ) L c ] + ( L m d + L c d )
where λ denotes the hyper-parameter to balance the whole loss, which is set to increase gradually from 0 to 1 as the iteration progresses.
Repeating the four steps of the above training process, the DCMSDA model is iteratively trained using the overall optimization objective. Through enough training, the feature extractors are able to extract domain-invariant fault features, thus allowing the shared fault classifier to accurately identify the fault type of target domain.

3. Experimental Verification

3.1. Experimental Dataset Description

To evaluate the effectiveness of the proposed bearing fault diagnosis method, the model, which is trained with labelled data, is expected to distinguish unlabeled data under different working conditions. The dataset provided by the University of Paderborn (PU) [20] is obtained from specially designed rolling bearings, including healthy state, inner raceway, and outer raceway faults. It contains 32 vibration signals, which can be spilit into healthy (6), artificially damaged (12), and real damaged (14), respectively, according to the accelerated lifetime, tested using 6203 grooved ball bearings [21]. Following [5], we only utilized the data on real damage. The modular test rig used to perform the accelerated life test is shown in Figure 4. The bearings used in the study were subjected to accelerated life by varying the rotational speed of the drive system, the radial force applied to the test bearings and the load torque on the driveline. In the test stage, a piezoelectric accelerometer was used to collect vibration signals from the bearing housing with a sampling frequency of 64 kHz.
The main operating parameters were the rotational speed, radial force, and load torque. Four diagnostic tasks are listed in Table 3. Thus, the domain adaptation task consisted of four different operating conditions. The proposed and compared methods were evaluated under 12 transfer tasks. For example, 0→1 denotes that the labelled data were taken as the source domain data, under the operation condition where task code 0 resides. On the contrary, the unlabeled data were utilized as target domain data.
In experiments, the bearings with actual damage were labelled at 0–12 for fault diagnosis tasks. Among the 13 faulty bearings, five failed in outer rings, five in inner rings, and the remaining failed in both inner and outer rings. Each fault category contained about 256,000 time-series sample points. With the segment length of 1024 data points, 250 samples per class were created, and each working condition contained about 250 × 13 samples. The four working conditions were composed of three working parameters, so the transfer difficulty was in line with the actual situation [22]. In addition, each fault category contained multiple fault parameters, such as main damage mode, damage degree and damage characteristics (the meaning of the specific content is explained in [20]). Therefore, the dataset for the experiments reflected potential variations of actual faults.

3.2. Training Details

The training parameter settings during the experiments are listed in Table 4. All experiments were conducted on a workstation as follows: CPU was Intel® Core™ I7-6850K@ 3.60 ghz, GPU was GTX1080Ti, video memory was 11 GB. The operating system was Windows 10, and the deep learning framework was Pytorch1.10.

3.3. Compared Methods Description

To demonstrate that our method performed better than other models, seven models were selected for comparison under the same dataset.

3.3.1. Domain Adaptation Based on Statistical Distance Metrics

According to [5], correlation alignment (CORAL) [23], MMD [24,25], MK-MMD, and Joint MMD (JMMD) [26] are often selected as statistical distance metrics. Here, we utilized the typical MK-MMD and JMMD as statistical distance metrics.

3.3.2. Domain Adaptation Based on Adversarial Learning

We adopted two commonly used methods, including domain adversarial neural network (DANN) [27] and conditional domain adversarial network (CDAN) [28], as comparative models.
In addition, we compared Resnet, a deep learning method without domain adaptation; MRAN [29], an earlier proposed method with extracting features on multiple spaces; and DDAN [30], a representative method with dynamic distribution adaptation. For a fair comparison, the experimental configurations and the dataset were consistent across all models. The methods and their corresponding transfer modules are shown in Table 5.

3.4. Results and Analysis of Comparative Experiments

3.4.1. Classification Accuracy

All methods under each transfer task were repeated 10 times with random initial parameters. Table 6 lists the average diagnostic accuracy of each method on the target domain. The average diagnosis accuracies of Resnet, DAN, JAN, DANN, CDAN, MRAN, DDAN, and the proposed DCMSDA were 50.31%, 69.44%, 72.38%, 73.51%, 73.79%, 71.25%, 71.34%, and 77.10%, respectively. In particular, for the simple transfer task of 2→0, the proposed model could still improve accuracy by 1.28%, compared with the CDAN model which already had an impressive accuracy. For the complex transfer task of 3→1, accuracies of the compared models were generally lower, while the proposed model obtained a high accuracy and outperformed the CDAN by 13%. All experimental results proved that our model could accurately classify faults in variable transfer tasks.
From Table 6 we can draw the following conclusions.
  • Comparison with Resnet:
Resnet does not use the domain adaption strategy, leading to significantly low accuracies under all tasks. The results indicate that for the scenario with variable working conditions, the classification performance of the model without domain adaptation may significantly reduce. However, our model overcomes this shortcoming.
2.
Comparison with MRAN:
Although MRAN adopts the multi-space strategy, it only aligns the distribution of the high-level fault features over multiple spaces, ignoring the general information of vibration signals contained in the low-level features. In contrast, our model proposes two feature extraction modules, thereby providing more comprehensive information of faults.
3.
Comparison with DDAN:
Although DDAN applies the dynamic adaptation strategy, it uses a linear classifier to calculate the adaptive coefficient, resulting in a poor fitting effect. Moreover, it only utilizes one feature extractor to extract features, which leads to singular fault features. In contrast, our model constructs non-linear domain discriminators to calculate the coefficient. Therefore, it is better matched to the actual working scenario.
4.
Comparison with JAN:
JAN is a method based on joint distribution adaptation, and it assumes that marginal and conditional distribution adaptations are equally important. Therefore, when faced with complex transfer scenarios, it is impossible to quantify the relative importance of aligning two distributions, leading to a poor diagnosis. In contrast, benefiting from the adaptive coefficient, our model can automatically select the appropriate distribution alignment.
5.
Comparison with CDAN:
CDAN achieves higher accuracy in the compared models. However, it only considers conditional distribution alignment. Moreover, this adversarial-based training approach usually faces the problem of slow convergence in model training. Thus, the diagnosis performance is inferior to our proposed method.

3.4.2. Accuracy Curves

To analyze the experimental results concisely, we chose the DDAN method with dynamic distribution adaptation, the JAN method with the best effect based on statistical distance metric, the CDAN method with the best effect based on adversarial learning, and our proposed model, as comparisons. Figure 5 shows the accuracy curves of DDAN, JAN, CDAN and the proposed DCMSDA model in the source domain training and the target domain test under the 2→0 transfer task, the optimal transfer task. There was a gap in accuracy between the source-domain training and target-domain test (Valid), which indicated that the distribution discrepancies between domains led to a lower generalization performance. Thanks to the proposed transfer modules, our model exhibited the smallest gap. This proved that the proposed model was easier to apply to actual working conditions. Additionally, in terms of the stability of the test (Valid) accuracy convergence, the proposed model significantly outperformed other methods. Figure 6 visually compares the test (Valid) accuracy curves for the four methods above. As can be seen, the accuracy of the proposed model was significantly higher than other methods.

3.4.3. Feature Visualization

To compare the domain alignment effects of different models, the fault classification results of DDAN, JAN, CDAN and the proposed DCMSDA model were visualized under the 2→0 transfer task using the t-distributed stochastic neighbor embedding (t-SNE) algorithm. As shown in Figure 7, two distinctive shapes indicate data from two domains, and diverse colors indicate varying fault categories. Some discussions of the feature representations are as follows:
  • The proposed DCMSDA model could obtain small intra-class distances and large inter-class distances, which suggests that our method has a strong fault classification capability. Specifically, as can be seen in Figure 7a, features in the DDAN method were somewhat jumbled with a poor gathering effect. Moreover, the category boundaries were not distinctly defined, which means that it was more difficult for the fault classifier to separate these features [31]. From Figure 7b,c, the JAN and CDAN methods incorrectly clustered the Source_IR12 fault and the Target_OR0 fault together, but the proposed method successfully separated them. From Figure 7d, the proposed method could achieve a better convergence effect of faults in the same category and obtain more obvious category boundaries;
  • The proposed DCMSDA model could extract representative domain-invariant features and exhibited excellent generalization performance because features of the source and target domains at the same fault category were closest. Specifically, as can be seen in Figure 7a–c, the three compared methods all closed the source domain and the target domain features of OR0, OR2 and IR12 faults unsuccessfully, but the proposed method closed them successfully, as shown in Figure 7d.

3.4.4. Confusion Matrix

The confusion matrices of DDAN, JAN, CDAN and the proposed DCMSDA model were compared under the 2→0 transfer task. Firstly, Figure 8a shows that DDAN had the lowest classification accuracy. The results in Figure 8b,c show that JAN had higher accuracy on certain label classification tasks, such as faults with label 1, and CDAN had higher accuracy on certain label classification tasks, such as label 9 and label 11. In contrast, the proposed model had higher accuracy than several other models on all label classification tasks, as shown in Figure 8d. Secondly, the results in Figure 8a–c show that when detecting the complex faults with label 7, diagnostic accuracies of the compared models were significantly lower, with only 73% for DDAN, 80% for JAN, and 76% for CDAN. In contrast, the proposed DCMSDA model could increase the accuracy to 90%, as shown in Figure 8d. All experiment results showed that the proposed model had excellent classification performance to identify variable types of faults.

3.4.5. Receiver Operating Characteristics (RoC) Curves and Area under Curve (AuC) Values

RoC curves of DDAN, JAN, CDAN and the proposed DCMSDA model were compared under the 2→0 transfer task, as shown in Figure 9. Here, we treated the Target_OR3 fault as the positive sample and the Target_OR4 fault as the negative sample. As can be seen, the proposed model had better classification performance because its RoC curve was closer to the coordinate point (0, 1), a point that only the perfect classifier would pass through. Moreover, we calculated the AuC value under each curve, as shown in the legend of Figure 9. The results showed that our model outperformed the other three models, as its AuC value was closer to 1.

3.5. Results and Analysis of Ablation Experiments

To verify the validity of the two feature extraction modules and the dynamic distribution adaptation module proposed in the model, we conducted experiments under the 3→1 transfer task with six cases. Specific cases and test accuracies are shown in Table 7.
The following conclusions can be drawn from Table 7.
  • Comparing cases 1 and 2 with the proposed model, two feature extraction modules focused on extracting informative features of vibration signals. The marginal feature extraction module could extract marginal features and obtain the general fault information. The multi-space conditional feature extraction module included convolution kernels of different depths and sizes, which could extract richer conditional features and obtain the information on fault categories, thereby guiding a more accurate result;
  • Comparing cases 3 and 4 with the proposed model, we adopted two different metrics to measure the distribution discrepancies, which contributed to exerting most of their respective strengths and guided the feature extraction modules to extract more diagnosis knowledge. MK-MMD focused on the global distribution and was suitable for aligning marginal features. LMMD was concerned with the relationship between two sub-domains within the same category, and was suitable for aligning conditional features;
  • Comparing case 5 with the proposed model, the adaptive coefficient dynamically measured the relative importance of marginal and conditional distribution alignments, thereby helping the model to adapt to complex cross-domain scenarios;
  • Comparing case 6 with the proposed model, the fault diagnosis model with domain adaptation aligned the distributions of domains, which significantly improved the robustness under the cross-condition diagnosis tasks;
  • Among them, experimental cases 3 and 5 with domain adaptation showed the largest reduction in accuracy compared with the proposed model. The results indicated that the strategy of adopting two metrics and the adaptive coefficient contributed the most to improving diagnostic accuracy.

4. Conclusions and Future Work

This paper presented a novel DCMSDA model based on domain adaptation for bearing fault diagnosis under cross-working conditions. It fully extracted domain-invariant features and achieved the alignment of the feature space. Firstly, the model considered the marginal distribution and conditional distribution discrepancies of the cross-domain vibration signals, as well as the applicability of different metrics. Therefore, two feature extraction modules were designed to extract the corresponding fault features separately, and two different metrics were adopted to align them. Then, an adaptive coefficient was employed to quantify the alignment proportion, so as to control the adaptation process. Through comparative experiments, the results demonstrated the superior performance of our method over current state-of-the-art methods, even in complex cross-domain scenarios. Furthermore, we verified the effectiveness of transfer modules in the proposed model through ablation experiments. The ablation results showed that the strategy of applying two metrics and the adaptive coefficient contributed the most to generalization performance.
Future work aims to find a more suitable multi-space integration strategy based on the variability of the fault features learned in each space, and to extend the proposed model to practical production applications, such as online fault diagnosis.

Author Contributions

Project administration, M.G.; Resources, Z.H.; Supervision, J.J.; Validation, C.L.; Visualization, Z.G.; Writing—original draft, X.Z.; Writing—review & editing, Z.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key Research and Development Program of China [Grant No. 2020YFB1710600], the Key Research Development Program of Zhejiang Province [Grant No. 2021C01111], and the National Natural Science Foundation of China [Grant No. 62171170].

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: http://groups.uni-paderborn.de/kat/BearingDataCenter/ (accessed on 3 July 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lu, W.; Liang, B.; Cheng, Y.; Meng, D.; Yang, J.; Zhang, T. Deep Model Based Domain Adaptation for Fault Diagnosis. IEEE Trans. Ind. Electron. 2016, 64, 2296–2305. [Google Scholar] [CrossRef]
  2. Luo, Z.; Wang, J.; Tang, R.; Wang, D. Research on vibration performance of the nonlinear combined support-flexible rotor system. Nonlinear Dyn. 2019, 98, 113–128. [Google Scholar] [CrossRef]
  3. Jia, S.; Wang, J.; Zhang, X.; Han, B. A Weighted Subdomain Adaptation Network for Partial Transfer Fault Diagnosis of Rotating Machinery. Entropy 2021, 23, 424. [Google Scholar] [CrossRef] [PubMed]
  4. Dong, Z.; Ji, X.; Zhou, G.; Gao, M.; Qi, D. Multimodal Neuromorphic Sensory-Processing System with Memristor Circuits for Smart Home Applications. IEEE Trans. Ind. Appl. 2022. [Google Scholar] [CrossRef]
  5. Zhao, Z.; Zhang, Q.; Yu, X.; Sun, C.; Wang, S.; Yan, R.; Chen, X. Applications of Unsupervised Deep Transfer Learning to Intelligent Fault Diagnosis: A Survey and Comparative Study. IEEE Trans. Instrum. Meas. 2021, 70, 3525828. [Google Scholar] [CrossRef]
  6. Zhiyi, H.; Haidong, S.; Xiang, Z.; Yu, Y.; He, Z. An intelligent fault diagnosis method for rotor-bearing system using small labeled infrared thermal images and enhanced CNN transferred from CAE. Adv. Eng. Inform. 2020, 46, 101150. [Google Scholar] [CrossRef]
  7. Wang, Y.; Han, M.; Liu, W. Rolling bearing fault diagnosis method based on stacked denoising autoencoder and convolutional neural network. In Proceedings of the 2019 International Conference on Quality, Reliability, Risk, Maintenance, and Safety Engineering (QR2MSE), Zhangjiajie, China, 6–9 August 2019; pp. 833–838. [Google Scholar]
  8. Shao, H.; Jiang, H.; Zhang, H.; Liang, T. Electric Locomotive Bearing Fault Diagnosis Using a Novel Convolutional Deep Belief Network. IEEE Trans. Ind. Electron. 2017, 65, 2727–2736. [Google Scholar] [CrossRef]
  9. Yan, R.; Shen, F.; Sun, C.; Chen, X. Knowledge Transfer for Rotary Machine Fault Diagnosis. IEEE Sens. J. 2019, 20, 8374–8393. [Google Scholar] [CrossRef]
  10. Othman, E.; Bazi, Y.; Melgani, F.; Alhichri, H.; Alajlan, N.; Zuair, M. Domain Adaptation Network for Cross-Scene Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4441–4456. [Google Scholar] [CrossRef]
  11. An, J.; Ai, P. Deep Domain Adaptation Model for Bearing Fault Diagnosis with Riemann Metric Correlation Alignment. Math. Probl. Eng. 2020, 2020, 4302184. [Google Scholar] [CrossRef]
  12. Zhu, J.; Chen, N.; Shen, C. A New Deep Transfer Learning Method for Bearing Fault Diagnosis Under Different Working Conditions. IEEE Sens. J. 2019, 20, 8394–8402. [Google Scholar] [CrossRef]
  13. Wan, L.; Li, Y.; Chen, K.; Gong, K.; Li, C. A novel deep convolution multi-adversarial domain adaptation model for rolling bearing fault diagnosis. Measurement 2022, 191, 110752. [Google Scholar] [CrossRef]
  14. Zhu, Y.; Zhuang, F.; Wang, J.; Ke, G.; Chen, J.; Bian, J.; Xiong, H.; He, Q. Deep subdomain adaptation network for image classification. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 1713–1722. [Google Scholar] [CrossRef]
  15. Hong, Y.; Hwang, U.; Yoo, J.; Yoon, S. How generative adversarial networks and their variants work: An overview. ACM Comput. Surv. (CSUR) 2019, 52, 1–43. [Google Scholar] [CrossRef]
  16. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? Adv. Neural Inf. Process. Syst. 2014, 27, 3320–3328. [Google Scholar]
  17. Ji, X.; Dong, Z.; Lai, C.S.; Qi, D. A Brain-Inspired In-Memory Computing System for Neuronal Communication via Memristive Circuits. IEEE Commun. Mag. 2022, 60, 100–106. [Google Scholar] [CrossRef]
  18. Gretton, A.; Borgwardt, K.; Rasch, M.; Schölkopf, B.; Smola, A. A kernel method for the two-sample-problem. Adv. Neural Inf. Process. Syst. 2006, 19, 513–520. [Google Scholar]
  19. Long, M.; Wang, J.; Cao, Y.; Sun, J.; Yu, P.S. Deep Learning of Transferable Representation for Scalable Domain Adaptation. IEEE Trans. Knowl. Data Eng. 2016, 28, 2027–2040. [Google Scholar] [CrossRef]
  20. Lessmeier, C.; Kimotho, J.K.; Zimmer, D.; Sextro, W. Condition monitoring of bearing damage in electromechanical drive systems by using motor current signals of electric motors: A benchmark data set for data-driven classification. In Proceedings of the PHM Society European Conference, Bilbao, Spain, 5–8 July 2016; Volume 3. [Google Scholar]
  21. Karpat, F.; Kalay, O.C.; Dirik, A.E.; Doğan, O.; Korcuklu, B.; Yüce, C. Convolutional neural networks based rolling bearing fault classification under variable operating conditions. In Proceedings of the 2021 International Conference on INnovations in Intelligent SysTems and Applications (INISTA), Kocaeli, Turkey, 25–27 August 2021; pp. 1–6. [Google Scholar]
  22. Dong, Z.; Qi, D.; He, Y.; Xu, Z.; Hu, X.; Duan, S. Easily Cascaded Memristor-CMOS Hybrid Circuit for High-Efficiency Boolean Logic Implementation. Int. J. Bifurc. Chaos 2018, 28, 1850149. [Google Scholar] [CrossRef]
  23. Sun, B.; Saenko, K. Deep coral: Correlation alignment for deep domain adaptation. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 443–450. [Google Scholar]
  24. Borgwardt, K.M.; Gretton, A.; Rasch, M.J.; Kriegel, H.-P.; Schölkopf, B.; Smola, A.J. Integrating structured biological data by kernel maximum mean discrepancy. Bioinformatics 2006, 22, e49–e57. [Google Scholar] [CrossRef]
  25. Sejdinovic, D.; Sriperumbudur, B.; Gretton, A.; Fukumizu, K. Equivalence of distance-based and RKHS-based statistics in hypothesis testing. Ann. Stat. 2013, 41, 2263–2291. [Google Scholar] [CrossRef]
  26. Long, M.; Zhu, H.; Wang, J.; Jordan, M.I. Deep transfer learning with joint adaptation networks. In Proceedings of the 34th International Conference on Machine Learning, PMLR 2017, Sydney, Australia, 6–11 August 2017; pp. 2208–2217. [Google Scholar]
  27. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-Adversarial Training of Neural Networks. J. Mach. Learn. Res. 2016, 17, 2030–2096. [Google Scholar]
  28. Long, M.; Cao, Z.; Wang, J.; Jordan, M.I. Conditional adversarial domain adaptation. Adv. Neural Inf. Process. Syst. 2018, 31, 1640–1650. [Google Scholar]
  29. Zhu, Y.; Zhuang, F.; Wang, J.; Chen, J.; Shi, Z.; Wu, W.; He, Q. Multi-representation adaptation network for cross-domain image classification. Neural Networks 2019, 119, 214–221. [Google Scholar] [CrossRef] [PubMed]
  30. Wang, J.; Chen, Y.; Feng, W.; Yu, H.; Huang, M.; Yang, Q. Transfer learning with dynamic distribution adaptation. ACM Trans. Intell. Syst. Technol. (TIST) 2020, 11, 1–25. [Google Scholar] [CrossRef]
  31. Naseer, S.; Hussain, W.; Khan, Y.D.; Rasool, N. Optimization of serine phosphorylation prediction in proteins by comparing human engineered features and deep representations. Anal. Biochem. 2020, 615, 114069. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Intelligent fault diagnosis: (a) without dynamic distribution adaptation; (b) with marginal distribution adaptation; (c) with conditional distribution adaptation; and (d) with dynamic distribution adaptation.
Figure 1. Intelligent fault diagnosis: (a) without dynamic distribution adaptation; (b) with marginal distribution adaptation; (c) with conditional distribution adaptation; and (d) with dynamic distribution adaptation.
Entropy 24 01122 g001
Figure 2. Framework of the proposed model.
Figure 2. Framework of the proposed model.
Entropy 24 01122 g002
Figure 3. Structure of the conditional feature extraction module.
Figure 3. Structure of the conditional feature extraction module.
Entropy 24 01122 g003
Figure 4. Experimental setup of PU: (a) electric motor; (b) torque measuring shaft; (c) rolling bearing test module; (d) flywheel; and (e) load motor.
Figure 4. Experimental setup of PU: (a) electric motor; (b) torque measuring shaft; (c) rolling bearing test module; (d) flywheel; and (e) load motor.
Entropy 24 01122 g004
Figure 5. Training accuracy of source domain and test (Valid) accuracy of target domain for different models under the 2→0 transfer task: (a) DDAN; (b) JAN; (c) CDAN; and (d) proposed.
Figure 5. Training accuracy of source domain and test (Valid) accuracy of target domain for different models under the 2→0 transfer task: (a) DDAN; (b) JAN; (c) CDAN; and (d) proposed.
Entropy 24 01122 g005
Figure 6. Test (Valid) accuracy of different models on target domain under the 2→0 transfer task.
Figure 6. Test (Valid) accuracy of different models on target domain under the 2→0 transfer task.
Entropy 24 01122 g006
Figure 7. Feature visualization for different models under the 2→0 transfer task: (a) DDAN; (b) JAN; (c) CDAN; and (d) proposed.
Figure 7. Feature visualization for different models under the 2→0 transfer task: (a) DDAN; (b) JAN; (c) CDAN; and (d) proposed.
Entropy 24 01122 g007
Figure 8. Confusion matrix for different models under the 2→0 task: (a) DDAN; (b) JAN; (c) CDAN; and (d) proposed.
Figure 8. Confusion matrix for different models under the 2→0 task: (a) DDAN; (b) JAN; (c) CDAN; and (d) proposed.
Entropy 24 01122 g008
Figure 9. RoC curves and AuC values of classifiers for different models under the 2→0 task.
Figure 9. RoC curves and AuC values of classifiers for different models under the 2→0 task.
Entropy 24 01122 g009
Table 1. Architecture of the conditional feature extraction module.
Table 1. Architecture of the conditional feature extraction module.
LayerKernel SizeKernel NumberStridesOutput Shape
CNN Block1(1, 1)641(32, 64)
Average pooling(7, 1)641(26, 64)
CNN Block2(1, 1)481(32, 48)
CNN Block3(5, 1)641(32, 64)
Average pooling(7, 1)641(26, 64)
CNN Block4(1, 1)641(32, 64)
CNN Block5(3, 1)961(32, 96)
CNN Block6(3, 1)961(32, 96)
Average pooling(7, 1)961(26, 96)
Table 2. Architectures of the marginal and conditional domain discriminators.
Table 2. Architectures of the marginal and conditional domain discriminators.
LayerOutput Shape
F1: Fully connected marginal domain discriminator layer(Batch_size, 1024)
F2: Fully connected marginal domain discriminator layer(Batch_size, 1024)
F3: Fully connected marginal domain discriminator layer with one sigmoid(Batch_size, 1)
F4: Fully connected conditional domain discriminator layer(Batch_size, 1024)
F5: Fully connected conditional domain discriminator layer(Batch_size, 1024)
F6: Fully connected conditional domain discriminator layer with one sigmoid(Batch_size, 1)
Table 3. The diagnostic tasks of PU dataset.
Table 3. The diagnostic tasks of PU dataset.
Task Code0123
Load torque (Nm)0.70.70.10.7
Radial force (N)100010001000400
Speed (rpm)150090015001500
Table 4. Training parameter settings during the experiments.
Table 4. Training parameter settings during the experiments.
ParameterValueParameterValue
Epochs300Sample length1024
Batch size64Marginal feature dimension512
Weight decay0.00001Fused conditional feature dimension256
Learning rate0.001--
Table 5. Methods and corresponding transfer modules used for fault diagnosis.
Table 5. Methods and corresponding transfer modules used for fault diagnosis.
MethodTransfer ModuleAdaptive Coefficient
ResnetNo transferNo
DANMK-MMDNo
JANJMMDNo
DANNAdversarialNo
CDANCondition-adversarialNo
MRANMulti-spaceNo
DDANMK-MMD and LMMDYes
ProposedMK-MMD, LMMD and Multi-spaceYes
Table 6. Means of the test accuracies in different tasks with the PU dataset (%).
Table 6. Means of the test accuracies in different tasks with the PU dataset (%).
Transfer TaskResnetDANJANDANNCDANMRANDDANProposed
0→124.2753.1362.9268.9368.1860.2558.1671.35
0→292.8394.3794.0393.2394.9293.5493.5695.51
0→352.1278.1483.3180.1285.0281.8782.4885.08
1→041.1357.6556.5060.1259.8356.5663.3273.63
1→245.2865.5769.8967.3466.7568.8967.1869.33
1→322.7437.1038.2443.4045.2237.2539.6146.49
2→090.9692.0394.4792.6095.0590.3390.6396.33
2→130.3557.7965.3468.0868.3164.5657.9168.66
2→359.0183.9988.3989.8288.9387.2386.8490.21
3→052.0781.1883.6881.8183.0181.3779.6683.06
3→134.5247.0044.6650.0943.0547.8751.2956.76
3→258.4285.3387.1086.5487.1985.3385.4388.77
Average50.3169.4472.3873.5173.7971.2571.3477.10
Table 7. Cases and test accuracies of ablation experiments.
Table 7. Cases and test accuracies of ablation experiments.
CaseTransfer ModuleAdaptive CoefficientTest Accuracy (%)
Case 1MK-MMD, LMMD and marginal feature extraction moduleYes55.21
Case 2MK-MMD, LMMD and conditional feature extraction moduleYes52.76
Case 3MK-MMD and two feature extraction modulesYes39.26
Case 4LMMD and two feature extraction modulesYes56.44
Case 5MK-MMD, LMMD and two feature extraction modulesNo50.46
Case 6Two feature extraction modulesNo30.54
ProposedMK-MMD, LMMD and two feature extraction modulesYes57.98
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zheng, X.; Gu, Z.; Liu, C.; Jiang, J.; He, Z.; Gao, M. Deep Transfer Network with Multi-Space Dynamic Distribution Adaptation for Bearing Fault Diagnosis. Entropy 2022, 24, 1122. https://doi.org/10.3390/e24081122

AMA Style

Zheng X, Gu Z, Liu C, Jiang J, He Z, Gao M. Deep Transfer Network with Multi-Space Dynamic Distribution Adaptation for Bearing Fault Diagnosis. Entropy. 2022; 24(8):1122. https://doi.org/10.3390/e24081122

Chicago/Turabian Style

Zheng, Xiaorong, Zhaojian Gu, Caiming Liu, Jiahao Jiang, Zhiwei He, and Mingyu Gao. 2022. "Deep Transfer Network with Multi-Space Dynamic Distribution Adaptation for Bearing Fault Diagnosis" Entropy 24, no. 8: 1122. https://doi.org/10.3390/e24081122

APA Style

Zheng, X., Gu, Z., Liu, C., Jiang, J., He, Z., & Gao, M. (2022). Deep Transfer Network with Multi-Space Dynamic Distribution Adaptation for Bearing Fault Diagnosis. Entropy, 24(8), 1122. https://doi.org/10.3390/e24081122

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop