Hui Zhang ,Huaji Zhou ,Li Wang and Feng Zhou,★
1The Ministry Key Laboratory of Electronic Information Countermeasure and Simulation,Xidian University,Xi’an,710071,China
2Science and Technology on Communication Information Security Control Laboratory,Jiaxing,314033,China
ABSTRACT This paper proposes a novel open set recognition method,the Spatial Distribution Feature Extraction Network(SDFEN),to address the problem of electromagnetic signal recognition in an open environment.The spatial distribution feature extraction layer in SDFEN replaces convolutional output neural networks with the spatial distribution features that focus more on inter-sample information by incorporating class center vectors.The designed hybrid loss function considers both intra-class distance and inter-class distance,thereby enhancing the similarity among samples of the same class and increasing the dissimilarity between samples of different classes during training.Consequently,this method allows unknown classes to occupy a larger space in the feature space.This reduces the possibility of overlap with known class samples and makes the boundaries between known and unknown samples more distinct.Additionally,the feature comparator threshold can be used to reject unknown samples.For signal open set recognition,seven methods,including the proposed method,are applied to two kinds of electromagnetic signal data:modulation signal and real-world emitter.The experimental results demonstrate that the proposed method outperforms the other six methods overall in a simulated open environment.Specifically,compared to the state-of-the-art Openmax method,the novel method achieves up to 8.87% and 5.25% higher micro-F-measures,respectively.
KEYWORDS Electromagnetic signal recognition;deep learning;feature extraction;open set recognition
Electromagnetic signal recognition (ESR) is a process of extracting features from a received signal and combining them with prior knowledge to identify the signal type or emitter.ESR,which includes modulation recognition and emitter recognition,plays an increasingly vital role in electronic reconnaissance.Traditional ESR relies on artificially designed extractors,and there are two methods:the likelihood-based approach[1-3]and the feature-based approach[4,5].The former calculates the likelihood function of the signal,estimates the unknown parameters,and then compares it with the preset threshold to obtain the signal type.The latter method extracts features from the signal and compares them with theoretical values.However,the above methods need artificial design extractors to extract effective features from electromagnetic signals.This requires not only domain expertise and complex parameter tuning but also complex and massive calculations.In addition,traditional feature extractors are often sensitive to signal variations and interferences.Traditional methods will perform poorly in harsh working environments such as those containing noise or signal changes,which will cause misjudgment and recognition errors.
Due to the increasing complexity of the electromagnetic environment,deep learning-based methodologies have become the dominant approach.Traditional ESR methods have several limitations,including difficulties in extracting features and inadequate adaptability to complex modulation signals.The powerful capabilities of deep learning,including feature learning and pattern recognition,can help address these limitations[6,7].In[8],a complex convolutional neural network was constructed for signal spectrum identification.A novel data-driven model based on Long Short-Term Memory(LSTM) was proposed in [9],which learns from training data without requiring expert features like higher-order cyclic moments.In[10],the authors designed an innovative network structure based on wavelet scattering convolution networks.In[11],authors used the adversarial losses of the feature layer and decision layer to generate adversarial examples with stronger attack performance.In addition,fewshot learning [12,13],semi-supervised learning [14-16],reinforcement learning [17] and lightweight models[18,19]for ESR were proposed,which were studied from the aspects of using a small amount of labeled training samples and reducing model complexity.However,they can only recognize the signal types seen in the training stage(known classes),which is called closed set recognition(CSR).In the case of signals that have never been seen before (unknown classes),CSR misclassifies them into known classes,which limits their application in the real world.
The real-world ESR task is usually limited by various objective factors,making it difficult to collect all signals required to train a classifier.Data quality and quantity play a key role in the development of deep learning models[20,21].Samples of unknown classes often lack clear labels and class information,rendering existing ESR methods ineffective.Modern electromagnetic environments pose many security challenges,including radio spectrum competition,interference,and spoofing.Identification of unknown categories of samples is essential for signal safety.Due to the inability of traditional methods to identify and classify unknown samples effectively in modern electromagnetic environments,signal security is becoming increasingly critical.For example,in spectrum safety monitoring,unallocated frequencies are not allowed.In electromagnetic equipment supervision,only electromagnetic equipment included in the safety range can be operated.Other equipment is identified as non-allocated and cannot be operated.In the above cases,collecting all types of samples is not practical and can be solved using open set recognition (OSR).OSR acquires incomplete knowledge in training,and unknown classes can be submitted to the algorithm at testing[22].Unlike CSR,OSR analyzes whether the sample type is known or not.OSR performs two tasks:classifying known class samples and rejecting unknown class samples.Specifically,OSR learns the features of each known class from the training setStraincontaining only known classes.For the testing setStest,which contains both known and unknown classes,OSR distinguishes known and unknown classes by comparing the sample features with the learned features of known classes,thereby rejecting unknown classes.
Traditional deep learning networks,such as convolutional neural networks (CNNs),use a Softmax classifier to address OSR.Softmax is a linear classifier that converts the output of multiple classifications into a probability distribution within the range[0,1]with a sum of 1.Because it divides the feature space based on the number of known classes,this classifier is insufficient for OSR tasks.Researchers explored two approaches to address this challenge.In the first approach,features derived from known class samples are used to predict unknown class samples,transforming the OSR problem into a CSR one.One notable method is the Openmax model[23]proposed by Bendale and Boult.This model assigns scores based on fitting a Weibull distribution and computes pseudo-activation vectors for unknown samples.Using Openmax,Prakhya et al.[24]explored open set text classification with promising results.Dang et al.[25,26]also incorporated extreme value theory into recognition methods to fit open set classifiers.To effectively model data distribution tails,these methods often require a solid understanding of probability statistics.However,they do not provide a principled approach to determining the tail size.Schlachter et al.[27] introduced a method that splits the given data into typical and atypical subsets using a closed set classifier.This enables the modeling of abnormal classes by leveraging atypical normal samples.However,determining the crucial hyperparameter split ratio for recognizing unknown class samples remains a challenge.
The second approach involves improving the learning methods or network frameworks of the models to enhance the consistency among samples from the same class and the distinctiveness between samples from different classes[28,29],thereby reducing the difficulty of recognizing unknown class samples.The more comprehensively known classes are acquired,the easier it is to reject unknown samples.Deep learning networks have a notable advantage as feature extractors since they can automatically extract features[30,31].Hassen et al.[32]proposed learning a neural network-based representation for OSR that incentivizes projecting class samples around the class mean directly.Using Siamese networks,Wu et al.[33] proposed an identification method for open set emitters that relies on classification loss,reconstruction loss,and clustering loss.Yue et al.[34] improved the learned spectral-spatial features by reconstructing the spectral and spatial features.Yang et al.[35]combined prototype-based classifiers with deep CNN for robust pattern classification.However,most methods neglect the distribution features of samples in the feature space,namely inter-class distribution and intra-class distribution,despite the adoption of some clever loss functions.Nevertheless,these models still depend on the outputs of fully connected layers for their outputs.
Therefore,this paper proposes a spatial distribution feature extraction network (SDFEN).The network utilizes the spatial distribution information of samples as the network output for the classifier.To enhance the performance of the network,most methods[32,36,37]do not directly use cross-entropy loss to train the model.However,these loss functions either do not consider both inter-class and intraclass distributions or are not conducive to convergence.To train the model,SDFEN introduces a hybrid loss function that improves the form of the loss function.Almost all existing OSR methods adopt threshold-based classification schemes from the perspective of discriminative models[24,38,39].A feature comparator compares depth features and thresholds to recognize unknowns.
The main contributions of this paper can be summarized as follows:
1) A spatial distribution feature extraction network,i.e.,SDFEN,is proposed for OSR of electromagnetic signals.SDFEN makes up for the lack of information between samples in traditional CNN.It uses the differences between samples as the main features extracted by the model,which is more suitable for recognition tasks.
2) To model spatial distribution,a hybrid loss creates intra-class compactness and inter-class separation for feature learning.This method enhances intra-class consistency and inter-class differences,improving sample categorization.
3) To distinguish between known and unknown classes,a feature comparator has been proposed which compares the spatial distribution feature with known class features.The feature comparator has been designed to effectively utilize spatial distribution information obtained from network outputs.This helps determine the boundaries between known and unknown classes in the feature space.
4) Statistically significant improvement is achieved by the SDFEN in modulation recognition and emitter recognition,with up to 8.87%and 5.25%higher micro-F-measures respectively when compared to the state-of-the-art Openmax method.
The rest of this article is organized as follows: Section 2 introduces the framework of the spatial distribution feature extraction network in detail.This section will elaborate on the specific implementation process of the spatial distribution feature extractor,the Euclidean distance classifier,and the feature comparator.Section 3 describes the experimental settings in detail.The experiments will be conducted in both open and closed environments.The results show that the proposed method has significant advantages in modulation recognition and emitter recognition.Finally,Section 4 summarizes this paper and discusses our future work.
As shown in Fig.1,SDFEN is composed of three main modules:the spatial distribution feature extractor(SDFE),the Euclidean distance classifier,and the feature comparator.The process involves mapping the original signal to a feature space using a spatial distribution feature extractor.Then,the Euclidean distance classifier predicts the type of signal,which belongs to a known class.The predicted label is then revised by comparing the spatial distribution feature with a threshold to obtain a corrected result.In addition,hybrid loss is used during training.
Figure 1:Framework of the spatial distribution feature extraction network(SDFEN)
SDFEN developed a spatial distribution feature extractor for extracting spatial distribution features from samples.By introducing specialized network structures and learning strategies,SDFEN enables the network to better extract and understand the spatial distribution relationship between samples,as well as to identify unknown samples.The spatial distribution feature extractor is improved based on CNN.Traditional CNN is usually composed of convolutional layers,pooling layers,and fully connected layers.As a result,convolutional layers and pooling layers are alternately applied,and then several fully connected layers are used to generate the final output.The convolutional layer uses the convolutional kernel to perform weighted summation on the input sample.CNNs can reduce training parameters by sharing weights across different local receptive fields.Traditional CNNs use the output after the full connection layer as the final feature and rely on the Softmax classifier for classification.As shown in Fig.2,the spatial distribution feature extractor adds a spatial distribution feature extraction layer(SDFEL)to the traditional CNN.The spatial distribution feature extractor consists of two main components: a CNN and a spatial distribution feature extraction layer.SDFEL maps the original features obtained from the CNN to the depth features in the feature space.By introducing the class center vector,SDFEL captures spatial distribution information from the input signal.It calculates the distance between the feature vector and the class center vector to provide valuable similarity information for subsequent classifiers.
Figure 2:Framework of the spatial distribution feature extractor
CNN extracts the original featuresf(x,θ) from the input signal.Here,xandθare the input data and CNN parameters,respectively.The SDFEL then takes the original features and processes them into depth features.To achieve this,the layer introduces a class center vector for each of theKknown classes.The distance between the feature vector and the class center vectors in the feature space provides valuable similarity information for subsequent classifiers.
The SDFEL operates as follows for a recognition task withKknown classes:1)The original center vectors of theKknown classes are randomly initialized,denoted asmifor classi.2)The distancesdibetweenmiandf(x,θ)are calculated,which is given by:
Here,d(·) is the Euclidean distance function.The resulting spatial distribution feature,denoted asf'(x,θ,m),is given by:
wherem=[m1,m2,···,mK] represents the class center matrix that will be updated during training using the spatial distribution feature extractor.3)During the training process,the class center vectormiis updated with the CNN parametersθ.
SDFEL maps the original features obtained from the CNN to the depth features in the feature space.By incorporating the class center vectors into the feature extraction process,the spatial distribution feature extractor can effectively capture the spatial distribution information of the input signals.This information is then used to create a feature vector rich in similarity information.This feature vector can be effectively used by the subsequent classifier.
The Euclidean distance classifier assesses the similarity between a signal and each known class based on the spatial distribution feature extracted by the spatial distribution feature extractor.The smaller the distance,the higher the similarity between the signal and the class.The probability that the signal samplexbelongs to the classiis inversely proportional to the distance betweenf(x,θ)andmi,which can be expressed as:
The prediction labelypredcan be expressed as:
Therefore,ypredcan be obtained directly from the spatial distribution extractorf'(x,θ,m).To ensure non-negativity and normalization of the probability,the expression of the Euclidean distance classifier is designed to be similar to the Softmax classifier.The probability that samplexbelongs to classiis given by:
Here,Kis the number of known classes.
To improve OSR performance,signal samples of the same class are expected to be as close as possible in the feature space.In contrast,signal samples of different classes should be as far apart as possible.To achieve this,a hybrid loss function,denoted asLHybrid,is designed to optimize SDFEN parameters.LHybridis composed of two parts: the cross-entropy loss based on Euclidean distance,denoted asLEDCE,and the Max-Min distance loss,calledLMax-Min.
For a samplexwith labely,LEDCEcan be calculated as follows:
Essentially,optimizingLEDCEmeans reducing the distance between the original feature vector ofxand the corresponding classycenter vector.LEDCEis concerned with minimizing the distance between the feature vectors of samples with the same label and the corresponding class center vectors.This helps ensure that samples of the same class are close to each other in the feature space,which better captures their intrinsic similarities.
LMax-Minconsists of the maximum intra-class distancedmax-intraand the minimum inter-class distancedmin-inter.The average distance between the classisamples and the classicenter vector is defined as the classiintra-class distance.The maximum value ofKintra-class distances is computed asdmax-intra,which is expressed as:
Therefore,following the principle of making the intra-class distance smaller and the inter-class distance larger,LMax-Minis designed as follows:
Finally,LHybridcan be expressed as:
whereλis a hyperparameter in the range 0 to 1.
In this section,a feature comparator is proposed to identify unknown classes and adjust predicted results.The feature comparator works by comparing a threshold valueτand an outlier scorescoreoutlierto determine the class of a given samplex.The corrected predicted resultforxis then obtained as follows:
whereK+1 represents the label of unknown classes.The outlier scorescoreoutlieris defined as the distance between the samplexand its closest class in the feature space.Specifically,it is computed as follows:
Obviously,the threshold is crucial for identifying unknown samples.As shown in Fig.3,the larger the threshold,the more likely it is for some unknown samples to be wrongly classified as known ones.Conversely,setting a smaller threshold increases the risk of identifying known samples as unknown ones.Therefore,the threshold selection determines OSR effectiveness to some extent.To setτ,the training set is used to calculate the average intra-class distancedavg-intraand the average inter-class distancedavg-interof the training samples.Similar to Eqs.(7)and(8),these values can be computed by:
wherexprepresents thep-th sample in the training set,andrepresents the corresponding class center vector.However,the actual values of inter-class distance and intra-class distance are not significant.What matters is their relative magnitude.Therefore,the spatial feature distribution measureois defined as the ratio ofdavg-intertodavg-intra,that is:
Figure 3:The influence of threshold on the recognition result
The feature distribution measureocan be used to represent the distribution of signal samples in the feature space.To account for different learning degrees of different signals from various models,distinct feature distribution measures are set to correspond to distinct threshold rules,as follows:
Two datasets of electromagnetic signals are used to evaluate modulation recognition and emitter recognition,respectively.
The experiments on modulation recognition will be carried out on the RADIOML 2016.10a dataset [40,41].RADIOML 2016.10a dataset is a synthetic dataset generated with GNU Radio.It consists of 11 modulations,including 8PSK,AM-DSB,AM-SSB,BPSK,CPFSK,GFSK,4PAM,16QAM,64QAM,QPSK,and WBFM.Each modulation is at varying signal-to-noise ratios(SNR),including -20,-18,-16,-14,-12,-10,-8,-6,-4,-2,0,2,4,6,8,10,12,14,16,and 18 dB.Each SNR of each modulation has 1000 signals,for a total of 220000 signals.The complex valued IQ is processed as 128×2.
Emitter recognition experiments will be conducted on the ACARS dataset.ACARS is a digital data link system for short message transmission between aircraft and ground stations.It is designed to monitor aircraft running status and provide data support for maintenance personnel.The ACARS dataset in this paper is collected by an antenna system installed on the top of a high building in Jiaxing,Zhejiang Province,China.The dataset contains 16000 signals belonging to 20 different aircraft evenly.The complex valued IQ is processed as 4096×2.
To approximate real-world scenarios,different levels of openness[42]have been simulated,which can be given by:
whereCTRis the number of classes in training,andCTEis the number of classes in testing.
The OSR experiment uses evaluation indicators such as acceptance recognition rate (ARR),rejection rate (RR),and micro-F-measure [38] to measure the performance of the model.ARR represents the proportion of samples that should be accepted and correctly classified,while RR represents the proportion of samples that should be rejected.These indicators measure the recognition accuracy of the model for known and unknown classes,respectively.ARR and RR can be expressed as follows:
whereTPi,TNi,FPi,andFNidenote the true positive,true negative,false positive,and false negative of classi,respectively.In addition,the micro-F-measure was used in the experiment to evaluate the overall recognition performance of the model.Micro-F-measure is defined as the adjusted average of micro-precision ratioPmicroand micro-recall ratioRmicro,which can be calculated by:
3.3.1Experiment1
In this particular experiment,the RADIOML 2016.10a dataset was used to simulate a closed set scene.To reduce the workload during training,certain SNRs were chosen,specifically-14,-12,-8,-4,-2,0,2,4,8,12,and 14 dB.For each SNR,75%of the signal samples were randomly selected as training samples,while the remaining were reserved for testing purposes.
In this study,a CNN was utilized as the primary feature extractor.Its structure is illustrated in Fig.4a,where Conv denotes a convolutional layer module.Each convolutional layer module contains a convolutional layer,a maximum pooling layer,and a batch normalization layer.The original features extracted from the CNN are fed into the spatial distribution feature extraction layer,followed by the hybrid loss function for network training.The batch size of the CNN was set to 100,and the number of iterations was set to 500.The optimizer used was stochastic gradient descent(SGD),with an initial learning rate of 0.05.The hyperparameter of(10)was set to 0.3.Four approaches were evaluated.The first approach(Softmax)used Softmax as the classifier and applied cross-entropy loss while training.The second approach(II-Loss+CE),as described in[32],used both II-Loss and cross-entropy loss as the training objective.The third approach(GCPL)incorporates prototypes into the model and adds prototype losses to train the network[36].The fourth approach(Hybrid)usedLHybrid.
Figure 4:The structures of the CNN for modulation recognition and emitter recognition
To enable a comparative analysis of the recognition accuracy of the four method sets,we present the confusion matrices for each method in Fig.5.Additionally,we illustrate the recognition accuracy of the four methods under different SNRs in Fig.6.The closed set recognition accuracy of the four methods are 61.99%,62.80%,63.01%,and 64.84%,respectively.As depicted in Fig.5,Softmax and II-Loss+CE show weak recognition abilities for 16QAM and WBFM signals.GCPL performs better than Softmax and II-Loss+CE,while Hybrid outperforms the others in recognizing these two types of signals.Moreover,Hybrid performs better in recognizing BPSK,GFSK,and 4PAM signals.As demonstrated in Fig.6,recognition accuracy is positively correlated with SNR.Overall,the four methods exhibit higher recognition performance at SNRs of 0,2,4,6,8,10,12,and 14 dB.In general,Hybrid has superior advantages over the other methods.
To validate the effectiveness of the proposed method in modeling distribution features,the distributions of signal samples under four methods in the feature space are illustrated in Fig.7.The performance is demonstrated using BPSK,GFSK,PAM4,QPSK,8PSK,and CPFSK with an SNR of 10 dB to avoid being influenced by poorly recognized signal types.The results show that II-Loss+CE,GCPL,and Hybrid,which utilize distance features as the loss function,exhibit stronger discriminative power in spatial distributions than Softmax,which uses traditional cross-entropy loss.In general,the proposed method demonstrates more compact intra-class distributions and more distant inter-class distributions.
Out of the four methods mentioned,Softmax directly relies on the output of the original features from the fully connected layer for classification.This is without considering the distribution of samples in the feature space.On the other hand,II-Loss,GCPL,and Hybrid take into account the distribution of samples in the feature space.However,GCPL does not consider inter-class distances while II-Loss includes intra-class distances but its loss function is not conducive to model convergence.Hybrid,unlike the other three methods,considers both inter-class and intra-class distances with a loss function designed in the form of a ratio to make the model easier to train.This method exhibits smaller intra-class distances and larger inter-class distances,which benefit the recognition and OSR task.By increasing the inter-class distance and decreasing the intra-class distance,more space in the feature space can be left for unknown samples and the difference between known and unknown classes can be more pronounced during classification.
Figure 5:Confusion matrices of four methods for modulation recognition
Figure 6:Comparison of modulation recognition accuracy with SNR
Figure 7:Spatial distribution feature of four modulation recognition methods
3.3.2Experiment2
An ablation study was carried out to verify the effectiveness of the spatial distribution feature extractor and the hybrid loss function in the proposed method.RADIOML 2016.10a dataset was used and signal samples with 10 dB SNR were chosen.Seventy-five percent of the signal samples were randomly chosen as training samples,and the rest were selected for testing.The specific structure and parameter settings of the model used in the experiments were the same as Experiment 1 in Section 3.3.1.The experimental results are shown in Table 1.As the results show,the deep learning model with the feature extractor has significantly higher recognition accuracy than the model without the feature extractor.In addition,training the model with the hybrid loss function resulted in better recognition performance.
Table 1: Ablation study results
3.3.3Experiment3
In this experiment,the open set scene was simulated on the RADIOML 2016.10a dataset and signal samples with an SNR of 10 dB were chosen.The number of unknown classes was set to 2,3,4,and 5,respectively,and the remaining classes were known.Similarly,75%of the signal samples of known classes were randomly selected as training samples,and the rest were used for testing.Therefore,the corresponding openness is 0.0513,0.0823,0.1181 and 0.1598.
Seven different methods were compared,excluding the four methods described in Section 3.3.1,Softmax,II-Loss+CE,GCPL,and Hybrid.Additionally,three more methods were added: Center[37],which utilized Center Loss for model training;CPL [36],which uses convolutional prototype learning to learn features;and Openmax,the state-of-the-art OSR method.In Table 2,the ARR,RR,and micro-F-measure of these methods in different openness is illustrated.The micro-F-measure in the table is abbreviated as Fmi.It is evident that in the open set scene,Hybrid performed overall better than the other methods in the classification of known classes and rejection of unknown classes.This is at the same openness level.When the openness is 0.1598,the proposed Hybrid method outperforms the state-of-the-art Openmax method by 8.87%in terms of the micro-F-measure.At the same time,the ARR is 7.76%higher and the RR is 9.24%higher.Combined with the CSR results in Section 3.3.1,it can be observed that methods with higher accuracy in CSR also exhibit better performance in OSR.Furthermore,there is significant variation in the recognition performance of the same method at different openness levels.However,as openness increases,recognition performance does not decrease.This can be understood by considering the inherent differences in class recognition performance due to variations among known classes,as demonstrated in Fig.5.Therefore,an increase in the number of known classes does not necessarily mean open set recognition becomes easier.
3.4.1Experiment1
In this experiment,the closed set scene was simulated on the Acars dataset.Seventy-five percent of the signal samples were randomly selected as training samples and the rest for testing.Fig.4b shows the CNN structure for emitter recognition.The batch size was set to 100,the number of iterations was set to 500,SGD was selected as the optimizer,the initial learning rate was adjusted to 0.05,and the hyperparameter of(10)was set to 0.3.Similarly,Softmax,II-Loss,and Hybrid were evaluated.
The closed set recognition accuracy of the four methods are 98.68%,98.20%,98.45%,and 99.18%,respectively.Similarly,the distributions of samples in the feature space under the four methods are shown in Fig.8.Compared with Softmax,GCPL,and II-Loss+CE,the proposed method has a larger average inter-class distance and a smaller average intra-class distance.Hybrid also achieves closed set recognition accuracy.
Table 2: Open set evaluation for modulation recognition
Figure 8:Spatial distribution feature of four emitter recognition methods
3.4.2Experiment2
In this experiment,the open set scene was simulated on the ACARS dataset and the number of unknown classes was set to 4,5,6,and 7,respectively.Then the corresponding openness is 0.0572,0.0742,0.0925,and 0.1124.Similarly,75% of the signal samples of known classes were randomly selected as training samples,and the rest were used for testing.
Similar to Section 3.3.3,seven methods are evaluated.The results are shown in Table 3.Obviously,in the open set scene,Hybrid performs overall better than the others in the classification of known classes and rejection of unknown classes.When the openness is 0.0572,the proposed method Hybrid outperforms the state-of-the-art Openmax method by 5.25%in terms of micro-F-measure.At the same time,the ARR is 7.38%higher and the RR is 3.03%higher.The Hybrid extracts spatial distribution features for classification,and direct feedback on the similarity between samples of the same class and the difference between samples of different classes into the model for training,which enables the model to directly model the distribution of samples in the feature space,so that similar samples are closer to each other and different samples are farther away.Therefore,unknown class samples can have a larger activity space.Let unknowns not overlap with known class samples.Combined with the experimental results in Section 3.4.1,it can also be inferred that methods that perform better on CSR have a positive effect on OSR.
Table 3: Open set evaluation for emitter recognition
It is true that the Hybrid has some advantages in the field of OSR.However,OSR experimental results show that the proposed method is sensitive to openness like most other methods.In essence,the model itself recognizes the differences between various types of samples,which affects recognition effects under different openness.Further research is needed on OSR which is more robust under different openness.
To address the recognition problem of electromagnetic signals in the open set environment and cope with the increasingly complex electromagnetic environment,this paper proposes a novel SDFEN.SDFEN introduces a dedicated network structure and learning strategy to better extract and understand spatial distribution relationships between samples.This supports the recognition of unknown class samples.The SDFEL in SDFEN is specifically designed to make up for the lack of information between samples in traditional CNN.It utilizes the differences between samples as the main features extracted by the model,which is more suitable for recognition tasks.To enhance intraclass consistency and inter-class differences,SDFEN designs a loss function that brings samples of the same class closer together in the feature space while increasing the distance between samples of different classes.This feature learning approach helps differentiate between different class samples and provides stronger discriminative capabilities.Additionally,SFDEN introduces a feature comparator to compare the features of samples with the learned features of known class samples.This distinguishes between known and unknown class samples.By comparing the distances or similarities between features,SFDEN can determine whether a sample belongs to a known class.The proposed method is validated in modulation recognition and emitter recognition tasks.The distribution plots of samples in the feature space under Softmax,GCPL,II-Loss+CE,and Hybrid provide evidence that the proposed method enhances the intra-class compactness of similar samples while increasing the separation between different class samples.Compared with other methods in simulated open set scenarios,the proposed method shows better overall performance.Specifically,in comparison to the state-of-the-art Openmax method,the proposed method achieves up to 8.87% and 5.25% higher micro-F-measures respectively.However,the experimental results also indicate that the proposed method is sensitive to openness,suggesting the need for further exploration of more robust OSR methods.
Acknowledgement:The authors wish to express their appreciation to the reviewers for their helpful suggestions which greatly improved the presentation of this paper.
Funding Statement:The authors received no specific funding for this study.
Author Contributions:The authors confirm contribution to the paper as follows:study conception and design: Hui Zhang,Huaji Zhou;data collection: Li Wang;analysis and interpretation of results: Li Wang,Feng Zhou;draft manuscript preparation: Hui Zhang,Feng Zhou.All authors reviewed the results and approved the final version of the manuscript.
Availability of Data and Materials:This paper uses two datasets.The first is the RADIOML 2016.10a dataset,which can be found at https://www.deepsig.ai/datasets.The second is the real-world ACARS dataset,which is collected by the antenna system installed on the top of a high building in Jiaxing,Zhejiang Province,China.Since this dataset is currently being used in many other studies,it is confidential.
Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.
Computer Modeling In Engineering&Sciences2024年4期