Contrastive Consistency and Attentive Complementarity for Deep Multi-View Subspace Clustering

2024-05-25 14:38JiaoWangBinWuandHongyingZhang
Computers Materials&Continua 2024年4期

Jiao Wang,Bin Wuand Hongying Zhang

School of Information Engineering,Southwest University of Science and Technology,Mianyang,621010,China

ABSTRACT Deep multi-view subspace clustering (DMVSC) based on self-expression has attracted increasing attention due to its outstanding performance and nonlinear application.However,most existing methods neglect that viewprivate meaningless information or noise may interfere with the learning of self-expression,which may lead to the degeneration of clustering performance.In this paper,we propose a novel framework of Contrastive Consistency and Attentive Complementarity (CCAC) for DMVsSC.CCAC aligns all the self-expressions of multiple views and fuses them based on their discrimination,so that it can effectively explore consistent and complementary information for achieving precise clustering.Specifically,the view-specific self-expression is learned by a selfexpression layer embedded into the auto-encoder network for each view.To guarantee consistency across views and reduce the effect of view-private information or noise,we align all the view-specific self-expressions by contrastive learning.The aligned self-expressions are assigned adaptive weights by channel attention mechanism according to their discrimination.Then they are fused by convolution kernel to obtain consensus self-expression with maximum complementarity of multiple views.Extensive experimental results on four benchmark datasets and one large-scale dataset of the CCAC method outperform other state-of-the-art methods,demonstrating its clustering effectiveness.

KEYWORDS Deep multi-view subspace clustering;contrastive learning;adaptive fusion;self-expression learning

1 Introduction

The high-dimensional and complex big data create new challenges to traditional clustering methods,such as hierarchical clustering[1],density clustering[2],and fuzzy clustering[3].In particular,the emergence of multi-view data has produced new requirements for clustering theory,models,and algorithms.Multi-view data are the different descriptions of the same object.Multi-view clustering(MVC) partitions the multi-view data by exploring consistent semantics with complementary information,which can improve clustering performance and the effectiveness of unsupervised data analysis[4,5].Existing MVC can be roughly categorized as matrix factorization-based [6,7],graph-based[8,9],subspace-based [10,11],and kernel-based methods [12,13].The multi-view subspace clustering(MVSC) methods have attracted more attention due to their easy implementation and outstanding performance[14,15].MVSC methods are based on the self-expression property that each sample data can be a linear or affine representation by others in the same subspace.They learn the consensus selfexpression to compute the affinity matrix of multi-view samples and then apply a spectral clustering algorithm to obtain clustering results.The traditional MVSC methods explore the intrinsic subspace self-expression by shallow model or linear embedding function.However,they can not be effectively applied in some nonlinear or complex scenarios attributed to their limited representation ability.

Recently,deep multi-view subspace clustering (DMVSC) methods have been proposed due to the powerful high-dimensional nonlinear representation ability [16,17].DMVSC methods project the high-dimensional nonlinear data into low-dimensional latent features by deep model and then learn the consensus self-expression of multi-view data based on feature space.The consensus selfexpression should reflect the consistent semantics and the complementarity across views.Most DMVSC methods learn the consensus self-expression by various fusion strategies at different stages,which can be divided into early fusion of the low-dimensional latent features and late fusion of the view-specific self-expressions,as shown in Fig.1.In the early fusion,the latent features of all views are concatenated or directly used to learn shared self-expression for exploring consistent semantics by different diversity constraints.In the late fusion,the self-expression of each view is learned from view’s latent features individually and then fused to get the consensus self-expression across views for exploring the maximum complementarity.The fusion strategies contain average value,weighted sum,concatenating,and attention mechanism-guided fusion.However,view-private meaningless information or noise in view data may disturb the learning of consensus self-expression[18]and further result in the degeneration of clustering performance.For example,face images with different details can describe the same person,where the details can be different facial expressions,decorations,and illumination conditions.Therefore,to reduce the effect of view-private meaningless information or noise,an alignment strategy should be designed to satisfy the requirement of consistency across views in DMVSC methods.

Figure 1: An illustration of fusion strategies in DMVSC methods

A novel framework of contrastive consistency and attentive complementarity (CCAC) for DMVSC is proposed to fully explore consistent semantics and complementary information across views by contrastive learning and attention mechanism,respectively.Specifically,we use an autoencoder to obtain the low-dimensional latent features of each view.View-specific self-expression is learned from these features by a fully connected layer embedded into the auto-encoder.Subsequently,all the view-specific self-expressions are aligned by contrastive learning to avoid the effect of viewprivate meaningless information or noise.The aligned view-specific self-expressions are assigned adaptive weights by attention mechanism and then fused by convolution kernel to achieve consensus self-expression for computing affinity matrix.Finally,a spectral clustering algorithm is conducted on the affinity matrix to cluster multi-view data.The main contribution of our CCAC method can be summed up as follows:

• We propose a contrastive alignment strategy based on multi-view self-expression space to avoid the effect of view-private meaningless information or noise on exploring consistent semantic information across views.Different from the previous contrastive alignment based on feature space in deep clustering,the contrastive alignment module is directly oriented to the clustering task,which can maximize the similarities of positive sample pairs while minimizing that of negative sample pairs.

• Based on the aligned view-specific self-expression,we introduce an adaptive fusion strategy guided by the channel attention mechanism and convolution kernel to learn consensus selfexpression of multi-view data,which can fully exploit the diverse complementary information across views for further boosting clustering performance.

• Extensive experimental results on four benchmark datasets and one large-scale dataset verify the effectiveness and excellent clustering performance of CCAC over other state-of-the-art methods.

2 Related Work

2.1 Deep Multi-View Subspace Clustering

DMVSC has attracted more attention because of the outstanding performance and nonlinear application,which focuses on learning a high-quality consensus subspace self-expression based on deep learning.Inspired by the typical deep subspace clustering network [19],the unsupervised deep auto-encoder is widely used to learn latent low-dimensional features of multi-view data.Then,consensus self-expression is learned from the latent features.According to the strategy of learning consensus self-expression,existing DMVSC methods can be roughly divided into two categories,i.e.,early fusion of the latent features and late fusion of the view-specific self-expressions.

In the first category,consensus self-expression is learned via a shared fully connected layer from the latent features with different diversity constraints.For example,different fusion functions,including sum,max-pooling,and concatenation,are proposed to combine the latent features of all the views for learning shared self-expression[20].The t-SNE is used to supervise the latent features for learning shared self-expression by considering the inter-view and intra-view distributions [21].The spectral clustering loss and classification loss are introduced and integrated into the deep auto-encoder network to supervise the shared self-expression learning while considering the impact of noise[22].Hadanard product is proposed to constrain discrimination of low-dimensional latent features extracted by autoencoder for learning shared self-expression with the global and local structure[16].The view-specific latent features are weighted by the self-attention mechanism and fused to learn shared self-expression by a fully connected layer[23].

In the second category,multiple fully connected layers are inserted into the auto-encoder to learn self-expressions for multiple views,which are further fused to obtain consensus self-expression.For instance,view-specific self-expressions are learned with the diversity constraint of the Hilbert Schmidt Independence Criterion and aligned with common self-expression that is used for clustering[24].All the view-specific self-expressions are fused to obtain the consensus self-expression by a fully connected layer[25].In[17],the view-specific self-expressions are weighted by the channel attention mechanism and fused by the convolution kernel to learn consensus self-expression with maximum complementarity for precise clustering.Information bottleneck is extended to explore view-specific information in the latent feature space.After obtaining the view-specific self-expressions,the average is regarded as the final result for clustering[26].

In the above methods for DMVSC,the early fusion strategy pursues consistent semantics,and the late fusion strategy focuses on exploring complementary information.The DMVSC methods can achieve superior clustering performance compared with traditional MVSC methods.However,they suffer from the following limitations:1)The view-private meaningless information or noise may interfere with the quality of consensus self-expression learning,which further leads to the degeneration of clustering performance;2) They cannot fully explore consistent semantics and complementary information across views in a unified optimization framework.

2.2 Contrastive Learning

Contrastive learning is a self-supervised representation learning for mapping similar samples close and dissimilar samples far apart in the latent feature space.Thus,contrastive learning can maximize the similarities of positive sample pairs and minimize that of negative sample pairs[27].It has been applied in existing deep MVC methods to explore consistency based on the latent feature space.The alignment of view-specific latent features based on contrastive learning has been demonstrated due to the excellent performance in deep multi-view clustering[18,28–30].DMVSC methods based on contrastive learning have been proposed for learning consensus self-expression.For example,contrastive learning is used to obtain the common feature representation for self-expression learning[31].The method can explore the consistency across views,which can not achieve the maximum complementarity to boost the clustering performance.

3 Proposed Method

To fully explore consistent semantics and complementary information across views,we propose a novel framework of contrastive consistency and attentive complementarity (CCAC) for DMVSC,which is implemented by a unified optimization network based on deep auto-encoder.The framework of our CCAC method is shown in Fig.2.It consists of the view-specific self-expression learning,contrastive alignment,and adaptive fusion modules.In this paper,multi-view data denoted asincludesNsamples of varying dimensionDvacrossVviews.

3.1 View-Specific Self-Expression Learning Module

The unlabeled multi-view data are usually represented by high-dimensional features accompanied by redundancy or random noise.As the unsupervised learning model for processing unlabeled data,auto-encoder is widely used to obtain the low-dimensional features with semantic information by filtering out the redundancy or random noise of original data.Therefore,we employ auto-encoder to map the original view data into the latent feature space by minimizing reconstruction loss.The auto-encoder is constructed by encoder and decoder with a symmetrical structure.Based on the latent feature space,existing DMVSC methods have two strategies to learn the consensus self-expression for clustering:1)Early fusion.From the fusing latent features,the shared self-expression is learned by a fully connected layer embedding into the auto-encoder;2)Late fusion.View-specific self-expressions are learned and then fused to obtain consensus self-expression.As analyzed above,the late fusion strategy learns diverse view-specific self-expressions for exploring complementary information,while the early fusion strategy only pursues consistency.Consequently,we adopt the late fusion strategy to learn the consensus self-expression.

Figure 2: The framework of our proposed CCAC method mainly consists of three modules:1)viewspecific subspace self-expression learning module,which is constructed by multiple auto-encoders to extract deep latent features of multi-view data and from them to learn view-specific self-expression coefficient matrix by inserting a fully connected layer into each auto-encoder pair;2) contrastive alignment module,which uses contrastive learning to align view-specific self-expression coefficient matrices of all views for achieving the consistency objective of multi-view data;and 3) adaptive fusion module,which comprehensively explores the complementary information of multi-view data by channel attentive mechanism guided fusion strategy

where the last term is the regularization loss error to make Cvwith block-diagonal structure by Frobenius norm for spectral clustering.The constraintdiag=0 can avoid the trivial solution for Cv,i.e.,all the samples are partitioned into the same clusters.

3.2 Contrastive Alignment Module

Since the self-expression coefficient matricesfor all views obtained by Eq.(1)may receive interference with the view-private meaningless information or noise,which shifts the self-expression relationship that reflects consistent semantics across views,resulting in the degeneration of clustering performance.Recently,contrastive learning has been widely applied to pursue the consistent semantic objective in deep multi-view clustering methods [32–34].They use contrastive learning to align the latent features of multiple views to explore consistent features.Different from the contrastive alignment based on the feature space,we use contrastive learning to align thebased on the self-expression space to achieve the consistency objective,with the following two considerations:1)The self-expression coefficient matrices describe the similar relationship of samples in multiple views,which are direct for clustering task;2)The consistent objective of self-expression coefficient matrices can boost exploring maximum complementary information across views.

Specifically,we select the view-specific self-expression coefficients of the same samples from different views as the positive pairs,and the self-expression coefficients of different samples from the same or different views as the negative pairs.For one sample in the self-expression coefficient matrixthere are(V-1) positive sample pairs andV(N-1) negative sample pairs.The objective of contrastive learning based on self-expression is to maximize the similarities of positive sample pairs while minimizing that of negative sample pairs.In our method,the similarities between the view-specific self-expression coefficient matrices Cvand Cpare measured by cosine distance:

where<·,·>is the dot product operator.The loss function of contrastive loss between arbitrarily two view-specific self-expression coefficient matrices can be formulated as[18]:

whereτis the temperature parameter.For theVviews,the accumulated view-specific self-expression contrastive loss across views can be denoted as:

Based on the view-specific self-expression contrastive loss across views,the self-expression coefficients of samples that belong to the same cluster will increase outstandingly,and oppositely,the coefficients of the samples in different classes will decrease.To this end,the consistent semantics across views can be mined,and the effect of view-private information or noise can be filtered out,which can boost clustering performance.

3.3 Adaptive Fusion Module

After aligning the self-expression coefficient matrices,they are fused to obtain the consensus selfexpression with the complementary information across multiple views.Existing DMVSC methods adopt average,weighted sum,or concatenation strategies to fuse the self-expression coefficient matrices.However,they neglect the discriminative relationship among the self-expressions and the different contributions to consensus self-expression varying by each coefficient matrix.Inspired by the work in[35],we use the channel attention mechanism to weigh the multiple self-expression coefficient matrices according to their discriminative contribution.Then the weighted matrices are fused by convolution kernel to obtain the final self-expression coefficient matrix for spectral clustering.

Subsequently,we use a convolution kernelkto fuse the scaled tensorCsfor obtaining the consensus self-expression coefficient matrix C of multi-view data,which can achieve the maximum complementarity across views.

Finally,the consensus self-expression coefficient matrix C is used to calculate the affinity matrix of multi-view data for obtaining the final clustering results.Therefore,the Cshould be constrained by the block-diagonal structure.As in Eq.(1),we use the Frobenius norm to form the regularization loss on C.

3.4 Loss Objective and Optimization

In consequence,the loss functions of the three modules,including view-specific self-expression learning(L1),contrastive alignment(L2),and adaptive fusion(L3),are integrated as the objective loss of our CCAC method,which is formulated as:

whereλ1andλ2are the hyperparameters for balancing the consistency and complementarity across views.

Algorithm 1 gives the optimization process of CCAC,including the pre-training and fine-tuning stages.In the pre-training stage,we use the lossto train the encoder and decoder for obtaining their initial parameters.In the fine-tuning stage,Eq.(8)is used to fine-tune the pre-trained encoder and decoder by embedding multiple self-expression learning layers for learning consensus self-expression C with contrastive alignment and adaptive fusion strategies.Finally,the affinity matrix is obtained by W=(|C|+|C|T)/2,and the clustering results are obtained by applying the spectral clustering algorithm on W.

4 Experiments

4.1 Experimental Setup

Datasets.We conduct experiments on four benchmark datasets: BBCSport1http://mlg.ucd.ie/datasets/segment.html,Extended Yale B2http://cvc.cs.yale.edu/cvc/projects/yalefacesB/yalefacesB.html,COIL203https://www.cs.columbia.edu/CAVE/software/softlib/,UCI digits4http://archive.ics.uci.edu/dataset/72/multiple+features,and one large-scale dataset:Caltech101-all5https://data.caltech.edu/records/mzrjq-6wc02.BBCSport depicts 5 topical sports news of 544 samples by two publishers.Extended Yale B consists of 38 people faces and each face takes 64 images under different illumination.The top 10 classes of three feature views are selected in our experiments.COIL20 contains 20 object images of 1440 samples with three feature views.UCI digits dataset shows the handwritten digit images of 2000 samples with three different views.Caltech101-all is a large-scale dataset,which consists of 101 object categories images and a background class with 9144 image samples.Each image is extracted with five image features to form different views.The sample data in the five datasets are reshaped into vectors.Table 1 summarizes the details of the five datasets.

Table 1: The details of the five datasets

Comparison methods.Our CCAC method is compared with the state-of-the-art multi-view clustering methods,including LRR[2013,TPAMI][36],DiMSC[2015,CVPR][37],LTMSC[2015,ICCV][38],AWP[2018,SIGKDD][39],t-SVD-MSC[2018,IJCV][40],ELTMSC[2019,TIP][41],SCMV-3DT [2019,TNNLS] [42],LMSC [2020,TPAMI] [43],MCLES [2020,AAAI] [44],TRPCA [2020,TPAMI][45],CDIMC-net[2020,IJCAI][46],DMSC-UDL[2021,TMM][16],SSSL-M[2022,TIP][47],MDMVSC[2023,ESWA][17],DeepMVC[2023,CVPR][30],and ESCC[2023,TKDE][48].

Implementation details.CCAC is implemented by PyTorch framework based on the code of DMSC-UDL,which is optimized by adaptive moment estimation optimizer.The encoder and decoder are constructed by a three-layer fully connected neural network with symmetrical structure.In the pretraining stage,the learning rate is set to 0.001 on five datasets.For the fine-tuning stage,it is set to 0.001 on BBCSport and COIL20,and 0.0001 on UCI digits,Extended Yale B,and Caltech101-all datasets.In the adaptive fusion module,the size of convolution kernel is set ask=3×3 to fuse the scaled viewspecific self-expression tensor for obtaining the consensus self-expression on all datasets.The optimal hyperparametersλ1andλ2are selected by grid search strategy in a range of{0.01,0.1,1,10,100}.The temperature parameterτis determined in a range of{0.1,0.2,0.3,0.4,0.5,0.6,0.7,0.8,0.9,1.0}.The detailed network structure and parameters are summarized in Table 2.

Table 2: The detailed network parameters of the CCAC method. Lr and T denote the learning rate and the number of epochs

Table 3: Clustering performance comparison on BBCSport dataset

Table 4: Clustering performance comparison on Extended Yale B dataset

Table 5: Clustering performance comparison on COIL20 dataset

Table 6: Clustering performance comparison on UCI digits dataset

Table 7: Clustering performance comparison on Caltech101-all dataset.The term-indicates that no available results have been reported

Evaluation metrics.Following the convention of clustering evaluation metrics [14,15],the six widely used performance metrics,including accuracy(ACC),normalized mutual information(NMI),adjusted rand index (ARI),F-score,precision,and recall,are used to fully evaluate the clustering performance and effectiveness of our CCAC method.

4.2 Experimental Comparative Result Analysis

Tables 3–7 give the comparative results with the 16 multi-view clustering methods on the five datasets in terms of the six evaluation metrics.The best performance results are highlighted in bold and the second in underline.To objectively exhibit the clustering performance of CCAC method,we perform experiments on large-scale Caltech101-all dataset 5 times and on other four datasets 30 times.The mean and standard deviation of clustering results are given in the tables.

From the tables,it can be observed that CCAC achieves better clustering performance than most state-of-the-art multi-view clustering methods.Specifically,the observations can be summarized as follows:

1.The clustering results of our CCAC method exceed other comparison methods on BBCSport,COIL20,UCI digits,and Caltech101-all datasets.Especially,On the COIL20 dataset,CCAC is superior to the suboptimal method ETLMSC,improving the clustering results by 3.9%,1.1%,3.7%,3.6%,5.2%,and 1.7%in terms of ACC,NMI,ARI,F-score,precision,and recall metrics.The clustering results of CCAC on the Extended Yale B dataset are close to the optimal clustering performance.It is one of the two methods that can achieve more than 99%clustering performance.

2.Compared with the deep MVC methods (SSSL-M,CDIMC-net,DeepMVC),our proposed CCAC method can obtain superior clustering results.CCAC integrates the advantages of deep learning and subspace learning to obtain a high-quality cross-view affinity matrix for clustering.Moreover,compared with the DMVSC methods(DMSC-UDL and MDMVSC),the CCAC method also performs well.For example,the clustering results of CCAC achieve above 98% on Extended Yale B,COIL20,and UCI digits datasets of all evaluation metrics.The outstanding clustering performance is attributed to the following observations.In DMSCUDL,the Hadanard product is used to constrain discrimination of low-dimensional features to learn shared self-expression with the global and local structure.In MDMVSC,all the viewspecific self-expressions are fused to obtain the consensus self-expression by channel attention mechanism and convolution kernel,which can explore the maximum complementarity of different views for precise clustering.However,it neglects the consistency across views for clustering.Compared with the two DMVSC methods,we use contrastive learning to align the view-specific self-expressions and channel attention mechanism to fuse the aligned selfexpressions,respectively.Therefore,CCAC can comprehensively explore the consistency and complementarity across views to improve clustering performance.

3.Our CCAC method can obtain the best clustering performance on large-scale Caltech101-all dataset among all methods.However,we can observe that the clustering results of all methods on this large-scale dataset are not as outstanding as other small-scale datasets.Moreover,the running time and memory space of CCAC and other subspace-based clustering methods increase with the number of multi-view data samples.It is because the subspace-based clustering methods learn the affinity matrix of multi-view and apply spectral clustering to obtain the final results,which need the running time ofOand memory space ofOat least.Therefore,we will optimize the CCAC method with learning-based anchors to improve efficiency and extend the application.

4.3 Model Analysis

Ablation studies.In this section,we perform ablation experiments to verify the effectiveness of contrastive alignment and adaptive fusion modules in CCAC.Taking the BBCSport dataset as an example,the clustering results of ablation experiments are given in Table 8.If the adaptive fusion module is not used,as in model A,we average the view-specific self-expression as consensus selfexpression.Compared to B with A,contrastive alignment of the view-specific self-expressions can improve clustering performance by 6.3%,4.4%,4.8%,6.6%,2.9%,and 4.7%in terms of ACC,NMI,AR,F-score,precision,and recall.In model C,we use adaptive fusion instead of average values to further boost clustering performance.According to models C and A,we can observe that the adaptive fusion module plays the most important role in performance improvement of 12.1%,21.4%,20.1%,18.2%,13.9%,and 16.0%.From the clustering performance improvement comparison by contrastive alignment and adaptive fusion,we can observe that adaptive fusion outperforms contrastive alignment for enhancing performance.This can be attributed to the following reasons.Multi-view data are the different descriptions of the same object,including consistent and complementary information.In this paper,contrastive alignment aims to obtain consistent information,and adaptive fusion can fully mine the complementary information.Compared with consistent information,complementary information is the main reason for multi-view learning working and an important factor in improving the performance of multi-view clustering.When the two modules work together,the CCAC method achieves outstanding clustering results.

Table 8: Ablation experiment results

Parameter sensitivity analysis.There are two hyperparametersλ1andλ2in our CCAC method to balance the contribution of consistency and complementarity for multi-view clustering.The parameter sensitivity experiments are performed to evaluate the effect of hyperparameters in the range of{0.01,0.1,1,10,100},which are implemented by the grid search strategy.The experimental results of NMI with respect toλ1andλ2on four datasets are shown in Fig.3.It is observed that CCAC method is insensitive to the two hyperparameters within a certain range,especially hyperparameterλ2.The NMI metric decreases on BBCSport whenλ1is from 0.01 to 0.1.It is affected on Extended Yale B whenλ1is from 10 to 100.The NMI metric of CCAC on the UCI digits and COIL20 datasets always keeps stable withλ1andλ2in the range of{0.01,0.1,1,10,100}.It further verifies the parameter insensitivity and effectiveness of our CCAC method.Furthermore,the effect of temperature parameterτin the contrastive alignment module on clustering performance is evaluated by experiments.Fig.4 shows the parameter sensitivity results of ACC on the four datasets.We can observe that the clustering results of CCAC are insensitive to the choice ofτon BBCSport,COIL20,and UCI digits.There is a slight fluctuation of ACC on Extended Yale B,but the clustering performance is stable when theτis from 0.3 to 0.6.

Figure 3: Parameter sensitivity analysis of λ1 and λ2 on the four datasets

Figure 4: Parameter sensitivity analysis of temperature τ on the four datasets

Convergence analysis.We analyze the convergence of CCAC by plotting the changes of loss function logarithmic value and evaluation metrics with epoch on four datasets,as shown in Fig.5.We can see the logarithmic value of loss monotonically descends until convergence,oppositely,the six clustering performance evaluation metrics increase rapidly and keep stable on the four datasets.For example,the convergence of CCAC on UCI digits is the fastest in the four datasets,which converges in 50 epochs with such stable clustering performance.Meanwhile,the clustering performance also increases at the highest rate on this dataset and remains stable over the epochs.CCAC on Extended Yale B shows the slowest convergence,and the clustering performance rises to the best values after 1600 epochs.

Figure 5: Convergence analysis on the four datasets

5 Conclusion

In this paper,we propose a novel framework of contrastive consistency and attentive complementarity for DMVSC.The auto-encoder is employed to learn the low-dimensional latent features of each view by reconstructing the original view data.Then,multiple self-expression learning layers are inserted into the auto-encoder to learn the view-specific self-expression.To fully explore the consistent semantics,all the view-specific self-expressions are aligned through contrastive learning to avoid the effect of view-private information.According to the diverse relationship,the aligned view-specific self-expressions are assigned adaptive weights and fused to obtain the consensus self-expression by channel attention mechanism and convolution operator,aiming at maximizing the complementarity of multiple views.Experimental results demonstrate the outstanding clustering performance and effectiveness of our method.

Acknowledgement:None.

Funding Statement:The authors received no specific funding for this study.

Author Contributions:Jiao Wang: Methodology,Software,Validation,Writing—original draft.Bin Wu:Supervision,Methodology.Hongying Zhang:Investigation,Writing—review&editing.

Availability of Data and Materials:Not applicable.

Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.