Complementary-Label Adversarial Domain Adaptation Fault Diagnosis Network under Time-Varying Rotational Speed and Weakly-Supervised Conditions

2024-05-25 14:40SiyuanLiuJinyingHuangJianchengMaLichengJingandYuxuanWang
Computers Materials&Continua 2024年4期

Siyuan Liu ,Jinying Huang ,Jiancheng Ma ,Licheng Jing and Yuxuan Wang

1School of Data Science and Technology,North University of China,Taiyuan,030051,China

2School of Mechanical Engineering,North University of China,Taiyuan,030051,China

ABSTRACT Recent research in cross-domain intelligence fault diagnosis of machinery still has some problems,such as relatively ideal speed conditions and sample conditions.In engineering practice,the rotational speed of the machine is often transient and time-varying,which makes the sample annotation increasingly expensive.Meanwhile,the number of samples collected from different health states is often unbalanced.To deal with the above challenges,a complementary-label(CL)adversarial domain adaptation fault diagnosis network(CLADAN)is proposed under time-varying rotational speed and weakly-supervised conditions.In the weakly supervised learning condition,machine prior information is used for sample annotation via cost-friendly complementary label learning.A diagnostic model learning strategy with discretized category probabilities is designed to avoid multi-peak distribution of prediction results.In adversarial training process,we developed virtual adversarial regularization(VAR)strategy,which further enhances the robustness of the model by adding adversarial perturbations in the target domain.Comparative experiments on two case studies validated the superior performance of the proposed method.

KEYWORDS Time-varying rotational speed;weakly-supervised;fault diagnosis;domain adaptation

1 Introduction

In recent years,condition monitoring has been extensively applied to anticipate and detect machinery failures.While the majority of current research in mechanical fault diagnosis predominantly assumes the constant speed,advanced methods have been proposed to mine fault features under time-varying speed conditions,including order tracking [1],stochastic resonance [2],and sparse representation [3],etc.However,these methods still rely on high quality information such as shaft speed,and the analysis process is extremely complicated and prone to problems such as signal distortion.Deep learning (DL) methods can automatically process condition monitoring data and feedback diagnosis results with limited prior knowledge or human intervention.With its powerful data nonlinear fitting capability,DL excels in a wide range of fault diagnosis benchmark tasks.Recently DL-based research has endeavored to mitigate the impact of time-varying rotational speeds.For example,Han et al.proposed the L1/2 regularized sparse filtering (L1/2-SF) model for fault diagnosis under large speed fluctuation [4].Huang et al.introduced a compensation technique to address nuisance information effects and therefore improve the robustness of the model under timevarying speed and variable load conditions[5].Despite these advancements,the learning process still falls under the category of supervised learning,which may not align with real industrial production characterized by weakly-supervised conditions,such as sample imbalance,limited labeled data,and sample insufficiency.Therefore,it is necessary to explore fault diagnosis techniques that accommodate both variable speeds and weakly supervised conditions.In this context,Yan et al.proposed a novel semi-supervised fault diagnosis method termed label propagation strategy and dynamic graph attention network(LPS-DGAT)[6].

Additionally,another common assumption in existing studies is the consistency in the distribution of training data and test data,which deviates from actual industrial conditions[7].To eliminate the negative impact of data distribution discrepancy on the accuracy of DL models,considerable attention has been directed towards domain adaptation(DA)[8].Particularly,unsupervised DA(UDA)approach is able to transfer strongly relevant knowledge from source domains with abundant labeled samples to target domains with lacking labels[9].Improving the generalization of models under non-ideal data conditions is typical of the UDA problem,which industrial site is currently facing.UDA methods can be roughly categorized as Discrepancy-based DA and Adversarial-based DA.Discrepancy-based DA[10,11]seeks aligned subspaces the source and target domains,which feature representation invariant(i.e.,reduces the distribution discrepancy between different data domains in a specific distance space).However,the Discrepancy-based DA approach which is computationally intensive,requiring sufficient data from both the source and target domains.Inspired by generative adversarial networks [12],the way of Adversarial-based DA [13–15] becomes an alternative.Adversarial-based DA can learn the metric loss function between domains implicitly and automatically,without explicitly entering a specific functional form.Qin et al.proposed a parameter sharing adversarial domain adaptation network (PSADAN),solving the task of unlabeled or less-labeled target domain fault classification[16].Dong et al.expanded cross-domain fault diagnosis framework with weakly-supervised conditional constraints,designing a dynamic domain adaptation model[17].Quite a number of worthwhile highlights are reflected in the above study.However,grounded in the perspective of engineering,an irreconcilable contradiction persists: Non-ideal data conditions and the model’s high standards for data quality.In the context of Big Data,the cost of collecting massive labeled dataset remains high even for source domains.In most cases,there may be only a few labeled data for each operating condition,and the remaining unlabeled data need to be analogized and inferred.Hence,reducing the cost of data annotation needs to be focused on.

In this paper,we propose a novel CL adversarial domain adaptation network(CLADAN)model to address the above-mentioned non-ideal conditions.The idea of CL was first applied to cross-domain fault diagnosis research,and we improved it.The model copes with both weakly supervised learning conditions and time-varying speed conditions,which are deemed to be tricky in the past.In addition,a regularization strategy is proposed to further improve the robustness of the model adversarial training process.More excitingly,the CL learning process can add true-labeled data for self-correction and updating of the model.The main contributions are summarized as follows:

(1) A less costly sample annotation method for domain adaptation is proposed.CL learning,integrated with adversarial domain adaptation method,is able to alleviate the effects of domain shift and labeled sample insufficiency in the source domain.As a cost-friendly auxiliary dataset,CL annotations improve the performance and accuracy of the prediction model.

(2)We propose a method of discretizing category probabilities to enable classifiers to make highly confident decisions.

(3)By adding perturbations to the samples involved in the adversarial learning,CLADAN forced domain discriminators to learn domain-invariant features independent of rotational speeds.

The remainder of the paper is organized as follows: Section 2 describes the proposed method.Experiments are carried out for validation and analysis in Section 3.Conclusions are drawn in Section 4.

2 Method Overview

2.1 Complementary-Label Learning

In the field of fault diagnosis,the success of UDA remains highly dependent on the scale of the true-labeled source data.Owing to the cost of acquiring massive true-labeled source data being incredibly high,UDA alone is hard to adapt to the weakly-supervised industrial scenarios.Fortunately,while determining the correct label from the many fault label candidates is laborious,choosing one of wrong label (i.e.,CL) is more available,such as annotating “outer race fault”as “not normal”,especially when faced with a large number of fault candidates.The CL corresponding to the true label is presented in Fig.1.Obviously,CL that only specifies the incorrect class of samples is less informative than the true label [18].However,it is practically difficult to accomplish unbiased estimation for all samples.Industrial sites often only collect ambiguous information when performing fault data annotation.At this point the true label no longer appears reliable,while the CL feedbacks information that is already available precisely.With the same cost control,we can procure more CL data than the true-label data.In contrast to traditional pseudo-label learning,the CL learning annotation process incorporates priori information in the field,including worker experience and maintenance manuals.This improves the authenticity and reliability of the description of CL.

Figure 1: True labels vs.CL

Below we formulate CL learning.LetX⊂Rddenote feature(input)space.Y:={y1,y2,...,yk}be a label(output)space,whereykis the one-hot vector for labelkfrom label space L={1,2,...,k}.When true labels exist,we usually assume that each example(x,y) ∈X×Y,wherexdenotes an instance andyis the true label corresponding tox,is independently sampled from an unknown data distribution with joint probability densityP(x,y).The training goal of a machine learning classifierf:X→Rkcan be represented as minimizing the follow risk:

Finally,a more versatile unbiased risk estimator[19]that is defined as:

According to[19],we used simultaneous optimization training utilizing a maximum operator and a gradient ascent strategy to avoid possible overfitting problems caused by-(k-1)L(f(x),y).

2.2 Virtual Adversarial Regularization

With low computational cost and diminished label dependence,Virtual Adversarial Training can measure local smoothness for givenP(x,y)of data.The output distribution of each input data point is trained to be isotropic smoothing by smoothing the model selectively in the most anisotropic direction[20].In contrast to transfer learning in the uniform speed condition,the time-varying speed condition leads to a domain-shift process that is also time-varying.Weak changes in external factors,such as rotational speed over time,constantly change the domain distribution of the data.To enhance the generalization ability of domain adversarial training under different speed conditions,identifying the direction of the perturbation that maximally affects the model output distribution is crucial.

Armed with this idea,we propose a regularization strategy for improving the robustness of the diagnostic model under the interference of speed fluctuations.The essence of adversarial domain adaptation is to find an adversarial directionradv,making it easier for the domain classifier to make a judgment error.However,radvis vulnerable to weak changes inGf(xi)(e.g.,the random effects of speed fluctuations onP(x|y)).Finally,VAR loss is given as:

ε>0 is a tuning parameter andε=2.5 in line with[21].Dis a function that measure distribution discrepancy andCis a constant.

2.3 Complementary-Label Adversarial Domain Adaptation Fault Diagnosis Network

Figure 2: Overview of the proposed methodology

Next,we inputxsfintoGyto compute the classification loss:

According to Eq.(3),the complementary label loss can be formulated as:

CLADAN training process can be summarized as Fig.3.For the cross-domain fault diagnosis problem,we combine conditional adversarial domain adaptation network [14] (CDAN) with CL learning.CDAN incorporates multi-linear conditioning that improving classification performance and entropy conditioning that ensuring transfer ability,respectively.In this study,CDAN will be applied to adversarial domain adaptation process.The optimization objective of CDAN is described as follows:

Figure 3: Architectures and training process of CLADAN

3 Experimental Verification

In this study,part of the hyperparameters were uniformly set:(1)batch size=16(2)epoch=100,start epoch=5(3)optimizer is SGD+momentum,weight decay=0.0005(4)ε=2(5)αis determined by the number of source domain true-label samples.(6)γ1=1e-4,γ2=5e-3(7)The length of the signals is 5000.The process of building the dataset in the experiment is shown in Fig.4.Previous studies have proposed multiple domain adaptation methods with superior performance,including Joint Adaptation Networks(JAN)[10],Deep Subdomain Adaptation Network(DSAN)[11],Domain Adversarial Neural Networks(DANN)[13],Conditional Adversarial Domain Adaptation(CDAN)[14],and Maximum Density Divergence (MDD) [15].DSAN is a discrepancy-based DA approach with better DA performance than CDAN due to capturing fine-grained subdomain characteristics.CDAN,which lacks VAR,can be seen as one of the ablation comparison experiments for CLADAN.The essence of MDD is to maximize intra-class density loss during DANN training,while ensuring domain confusion and domain alignment.In the following subsections,we use these methods for comparison with CLADAN,with all related parameters remaining consistent.

Figure 4: Dataset construction process

3.1 Ottawa Bearing Data

Ottawa bearing data[24]was obtained from a single sensor under time-varying speed conditions.Two ER16K ball bearings are installed,and one of them is used for bearing fault simulation experiments.The health conditions of the bearing include (1) healthy,(2) faulty with an inner race defect,(3)faulty with an outer race defect,(4)faulty with a ball defect,and(5)faulty with combined defects on the inner race,the outer race and a ball(Serial number Corresponding to sample labels 0–4).

All these data are sampled at 200 kHz and the sampling duration is 10 s.The Ottawa dataset contains four variable speed conditions: (i) Increasing speed (IS);(ii) Decreasing speed (DS);(iii)Increasing then decreasing speed(ID);(iv)Decreasing then increasing speed(DI).

Each variable speed condition has data files for three different speed fluctuation intervals.Details are shown in Table 1.Therefore,we construct 12 data domainsOtfor experimental validation between different domains,wherendenotes variable speed condition(i–iv)andmdenotes three different speed fluctuation intervals.We sampled the raw data in overlap with a step size of 500,retaining 3000 normal samples and 1500 randomly selected fault samples.Random selection of fault samples can further satisfy the sample imbalance condition.

Table 1: Tasks descriptions of Ottawa dataset

CLADAN training process in task T7_T12(source=T7,target=T12)is shown in Fig.5,where CL0,CL400,and CL800 indicate that 0,400 and 800 true-labeled samples are input to the algorithm model,respectively.The rest of the training samples were labeled by CL,i.e.,the proportion of truelabeled samples was set to 0,10%,and 20%,respectively.After 100 epochs of training,the accuracy fluctuations of CLADAN in the source and target domains are shown in Fig.5.As true-labeled samples increase,the fluctuation ranges of classifier accuracy decreases.When the proportion of real samples is low,maintaining model accuracy becomes challenging at a steady state.This phenomenon indicates that the adversarial learning process is not stable enough due to insufficient true-labeled samples.Besides,CLADAN still maintains some performance with a final convergence accuracy of 60% when true-labeled samples are not available.Fig.6 visualizes the output distribution for the unlabeled samples in the target domain.

Figure 5: CLADAN training process in task T7_T12(a)source(b)target

Figure 6: Visualization of target domain features in task T7_T12

The full comparative results of the experiment are shown in Fig.7,and accuracy values are derived from the average accuracy of the last 10 epochs of the training process.By analysis,the performance of discrepancy-based DA is significantly weaker than that of adversarial-based DA under weakly-supervised conditions.CL contain limited information content such that domain confusion methods based on distance metrics are hard to blur the boundary sufficiently between the source and target domains.Adversarial-based DA reinforces the mutually exclusive role of complementary labels by deceiving domain discriminators.In addition,the performance of domain adaptation is further enhanced by assigning different weight coefficients to different weakly supervised samples by entropy conditions in CLADAN and CDAN.During the adversarial process,MDD executes domain alignment from the feature space rather than the original data space,realizing the learning of domain invariant features.It was found that while MDD also mitigates the disturbance of time-varying speed fluctuations,it leads to delayed convergence of the training process.In Fig.7b,CLADAN consistently demonstrates excellent diagnostic capability in target domain under different data conditions.The accuracy of CLADAN is still not high enough in the absence of sufficient true-labeled samples due to the model cannot be trained sufficiently by CL samples with insufficient information.Based on the above results a reasonable hypothesis can be made that when the sample length increased or the number of sensor channels increased,the information content per unit time corresponding to CL will be expanded.

Figure 7: Accuracy comparisons in Ottawa(a)radar diagram in CL400.(b)Task T7_T12

3.2 HFXZ-I Wind Power System Simulation Experimental Platform

The development of wind power systems is in line with the current international trend of“reducing carbon emissions,achieving carbon neutrality”.The HFXZ-I in Fig.8 is currently a simulation experiment platform for common wind power transmission systems.The relevant basic parameters are shown in Table 2.We obtained 11 channels and 4 channels of raw data on the planetary gearbox and helical gearbox,respectively,where two three-axis acceleration sensors were installed on the planetary gearbox to obtain the dynamic response in different vibration directions.All these data are sampled at 10.24 kHz and the sampling duration is 60 s.The gearbox health status details and task details are shown in Table 2.We set four variable speed conditions: (i) Decreasing speed 50–0 Hz,0.5 HP(ii) Increasing speed 0–50 Hz,0.5 HP (iii) Decreasing speed 50–0 Hz,1 HP (iv) Increasing speed 0–50 Hz,1HP.The data step size of overlapping sampling is 500.We obtained four data domains,each containing 2000 normal samples and 1000 faulty samples.We have fewer samples compared to Ottawa bearing experiments,but each sample has more abundant information due to the additional monitoring channels.

Table 2: Tasks descriptions of HFXZ-I dataset

Figure 8: HFXZ-I wind gearbox experiment platform

CLADAN training process in task T2_T1 (source=T2,target=T1) is shown in Fig.9,where CL0,CL200,CL400 indicate that 0,200 and 400 true-labeled samples are input to the algorithm model,respectively.The proportion of true-labeled samples in the source domain still satisfies the dataset construction condition.Meanwhile,some of the fault types being set,such as gear wear,have much weaker signal feedback due to the minor degree of the fault,meaning that the learning process of the model is more difficult.After 100 epochs of training,the accuracy fluctuations of CLADAN in the source and target domains are shown in Fig.9.It can be found that when the proportion of true-labeled samples reaches 10%(i.e.,CL200),the training accuracy of the source domain is already close to CL400(20%true-labeled samples).Compared with Ottawa bearing experiments,the target domain’s overall accuracy also has a significant improvement,while the accuracy fluctuation tends to moderate.This suggests that as the amount of information in true-labeled data increases,CL learning performs better and CLADAN is more stable.Fig.10 visualizes the output distribution for the unlabeled samples from HFXZ-I dataset in the task T2_T1 target domain.It can be intuitively seen that the inter-class distance in the target domain is amplified and the intra-class distance is reduced.This suggests that CL learning that fuse multiple sources of information are able to retain a greater amount of information,guiding CLADAN to better identify unlabeled fault samples under weakly supervised conditions.All fault diagnosis comparison experiments are shown in Fig.11.The model performance improves significantly at CL200.

Figure 9: CLADAN training process in task T2_T1(a)source(b)target

Figure 10: Visualization of target domain features in task T2_T1

Figure 11: Accuracy comparisons in HFXZ-I(a)radar diagram in CL200.(b)Task T2_T1

3.3 VAR Ablation Study and Parameter Optimization

In this subsection,we conduct ablation experiments to show the contribution of the different components in CLADAN.In particular,it is necessary to further verify the immunity of the VAR to speed fluctuations.Since VAR is implicit in the training process,it is thus difficult to interpret its suppression of time-varying speed intuitively from a signal perspective.However,external conditions,such as time-varying rotational speeds,can lead to reduced generalization of cross-domain diagnostic models,while it is undoubtedly true that adding local perturbations to adversarial training can enhance model generalization.Therefore,we reasonably believe that VAR will improve the robustness of the model and thus suppress the time-varying speed disturbance implicitly.DANN and CDAN can be considered as two sets of ablation experiments.Based on the gearbox experimental data CL400,we consider following baselines: (1) DANN: Train CLADAN without conditioning and VAR,namely train domain discriminatorGdonly based on the output features ofGf,VAR and the predicted outer product ofGydo not participate in training.(2)CDAN:Train CLADAN without VAR.(3)WO/D:Train CLADAN without the discretizing category probabilities in Eq.(11).The results of the ablation experiment are shown on the left in Fig.12.It can be found that adding conditional distributions on the input side captures the multimodal structures of distributions with cross-covariance dependency between the features and classes,improving transfer learning performance significantly.Comparing CLADAN with WO/D was able to demonstrate that discretizing category probabilities has a positive effect on domain adversarial training for the CL dataset.Discretizing category probabilities can sharpen the representation of conditional probabilities when CL is used as an input to the conditional distribution.While CLADAN’s performance is not guaranteed to be optimal at all times during transfer tasks,its overall fault diagnosis performance is still superior.In addition,we perform an optimization search for the key parameters in Eq.(13)with results in Fig.12 right.It can be observed that CLADAN performs best forα=0.3 andλ=1.2.

Figure 12: Results of ablation study and parameter optimization

Current DA-based fault diagnosis techniques often require high-quality and sufficient sourcedomain labeled data.However,a large number of data samples with pending labels in the industrial field are is unavailable for direct training,resulting in the underutilization of these feature-rich data resources.Hence,the CL-based weakly supervised learning approach can be employed to assign fuzzy conceptual labels to these data samples.Meanwhile,CL differs from traditional pseudo-labeling methods in that its annotation process leverages a priori knowledge,yet its annotation cost is less than that of labeling real samples.The efficacy of DA relies not on labels but on generalizing the generalized feature extraction patterns by capturing the underlying structure of the training data.Thus,in domain-adversarial training,injecting random perturbations into the target domain can also substantially improve the robustness of the cross-domain fault diagnosis model,leading to better learning of domain-invariant features under time-varying rotational speed conditions and implicitly removing the interference of fault-irrelevant components in the data content.

4 Conclusions

In this paper,CLADAN is developed for typical non-ideal scenarios of industrial sites.We integrate time-varying rotational speed conditions and weakly supervised learning conditions into cross-domain fault diagnosis,and use budget-friendly complementary labels to annotate unlabeled data in the source domain.In experiments,we establish a series of demanding conditions and complex faults to simulate industrial scenarios and we obtain accurate diagnostic results with CLADAN.The subsequent steps will delve into domain adaptation techniques for multiple CLs of the same sample,enhancing the learning performance of CL0,(3)Alternative solutions for time-varying speed conditions.

Acknowledgement:The authors wish to express their appreciation to the reviewers for their helpful suggestions which greatly improved the presentation of this paper.

Funding Statement:This work is supported by Shanxi Scholarship Council of China(2022-141)and Fundamental Research Program of Shanxi Province(202203021211096).

Author Contributions:The authors confirm contribution to the paper as follows: Study conception and design:Liu S.,Wang Y.and Huang J.;methodology:Liu S.;software:Liu S.;validation:Liu S.,Huang J.and Ma J.;formal analysis:Liu S.;investigation:Jing L.;resources:Ma J.;data curation:Liu S.;writing—original draft preparation:Liu S.;writing—review and editing:Liu S.;visualization:Liu S.;supervision:Huang J.;project administration:Huang J.;funding acquisition:Huang J.and Liu S.All authors have read and agreed to the published version of the manuscript.

Availability of Data and Materials:Data available on request from the authors.The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.