Fengyong Li,Weicheng Shen,Zhongqin Bi and Xiangjing Su
1College of Computer Science and Technology,Shanghai University of Electric Power,Shanghai,201306,China
2College of Electrical Engineering,Shanghai University of Electric Power,Shanghai,201306,China
ABSTRACT False data injection attack(FDIA) is an attack that affects the stability of grid cyber-physical system (GCPS) by evading the detecting mechanism of bad data.Existing FDIA detection methods usually employ complex neural network models to detect FDIA attacks.However,they overlook the fact that FDIA attack samples at public-private network edges are extremely sparse, making it difficult for neural network models to obtain sufficient samples to construct a robust detection model.To address this problem,this paper designs an efficient sample generative adversarial model of FDIA attack in public-private network edge,which can effectively bypass the detection model to threaten the power grid system.A generative adversarial network (GAN) framework is first constructed by combining residual networks(ResNet)with fully connected networks(FCN).Then,a sparse adversarial learning model is built by integrating the time-aligned data and normal data, which is used to learn the distribution characteristics between normal data and attack data through iterative confrontation.Furthermore,we introduce a Gaussian hybrid distribution matrix by aggregating the network structure of attack data characteristics and normal data characteristics, which can connect and calculate FDIA data with normal characteristics.Finally, efficient FDIA attack samples can be sequentially generated through interactive adversarial learning.Extensive simulation experiments are conducted with IEEE 14-bus and IEEE 118-bus system data,and the results demonstrate that the generated attack samples of the proposed model can present superior performance compared to state-of-the-art models in terms of attack strength,robustness,and covert capability.
KEYWORDS Distributed smart grid;FDIA;adversarial learning;power public-private network edge
With the continuous advancement of smart grid construction, the traditional power system is deeply integrated with information control equipment and communication sensor networks to form the grid cyber-physical system (CPS) [1].The new energy internet contains a high proportion of distributed new energy, which is a large-scale power CPS that can more effectively leverage the advantages brought by information fusion.However,the CPS used in smart grids and other Industrial Internet of Things (IIoT) networks face huge security challenges, especially in today’s increasingly serious threat of network attacks,and are also more susceptible to network attacks.Network attacks will not only damage the normal functions of the information system,but also may be transmitted to the physical system,threatening the safe operation of the power system[2].Tian et al.[3]conducted in-depth research on the unique security issues surrounding safety critical CPSs.They proposed two adversarial attack methods based on regression models and produced almost imperceptible adversarial images, which pose a significant threat to the navigation and control of unmanned aerial vehicles(UAV).State estimation is the calculation of the state of the power grid system based on the raw measurement values collected by supervisory control and data acquisition(SCADA),and plays a very important role in the control center.In particular,damaged system state estimation may interfere with the operation of the power system,as many power system applications are dependent on the results of state estimation[4].Among prevalent network attacks,FDIA represents a common attack vector that compromises the integrity of power grid information systems.Its main goal is to tamper with power grid measurement data through false attack data injection to damage power system state estimation[5],thereby causing line congestion[6],power outage[7],communication congestion,etc.It is one of the most threatening attack methods to the power system.
In recent years,there have been multiple large-scale power network security incidents worldwide,including scenes of using data tampering mechanisms for attacks.Taking the attacks on Ukrenergo as an example, Ukrenergo suffered two network attacks in 2015 and 2016, where attackers injected false data and deleted original data into the SCADA system, resulting in huge economic losses and social unrest[8].These events indicate that attackers will bypass bad data detection(BDD)through carefully designed fake data, and their network attacks will pose a huge threat to the power grid.Therefore,a large amount of research work has been carried out in society to defend against FDIA.For example,Luo et al.[9]used adaptive kalman filter to estimate the internal state and noise of the system.Cheng et al.[10]introduced a highly discriminative FDIA detector known as thek-smallest residual similarity (kSRS) test.This method achieves high detection rates and low false alarm rates for FDIA by portraying the statistical consistency of measurement residuals in AC state estimation.In addition,FDIA detection methods based on artificial intelligence have been proposed,such as support vector machine(SVM)[11],convolutional neural network(CNN)[12],random forest(RF)[13],etc.This type of method does not require pre-acquisition of power system model parameter information,possesses substantial computing power,can detect attacks quickly and on a large scale,and can achieve some good results.However,the above methods rely on cloud computing technology,and all collected data is processed and applied in the data center of the power grid.Limited communication and storage resources make the power grid lack the ability to process data in real-time,which can result in some key operations of the power grid not being executed in a timely manner.Therefore,to further enhance attack sample impact,real-time sample generation at the public-private network edge is necessitated,circumventing the existing centralized FDIA detection.
Furthermore, in order to better predict potential FDIA attacks in the power public-private network edge,domestic and foreign researchers have developed various studies to investigate possible methods for constructing FDIA[14,15].On the one hand,some literature focuses on the direction of constructing false data injection attack vectors.For example,a continuous FDIA[16]was proposed to increase the attack time span to a certain period and identify the sequence of lines that need to be heavily protected.Liu et al.[17]designed a sparse attack strategy that is also applicable to situations where topology information is incomplete.Tian et al.[18]introduced adversarial examples and FDIAs(AFDIAs) to study state estimation attack scenarios in power systems.By adding perturbations to state variables instead of measured values, the attack can be carried out covertly and avoid being detected by BDD.This method optimizes the final effect of the attack.On the other hand, due to the relative closeness of the private network side of the virtual power plant (VPP), the information flow attack behavior of the public network in the actual scene is difficult to easily cross the protection of the public and private boundaries, and the attack samples are extremely rare or even difficult to obtain.The emergence of some sample construction methods based on artificial intelligence provides new ideas.For example, Ehrhart et al.[19] designed a GAN framework combining long short-term memory(LSTM)network with fully convolutional network(FCN),which can generate the training data set of time series better.Rao et al.[20]proposed an intrusion data enhancement method based on the combination of generative adversarial network and deep neural network, which generated highquality false data injection(FDI)attack samples and solved the problem of insufficient attack samples in the virtual power plant scene.Gao et al.[21] utilized GANs’ability to learn features from input data and proposed GAN-based automatic property generation (GAPG) to automatically generate verification properties that support model checking.This innovative approach shows promise in automating property generation for model checking,reducing the need for manual specification.Dina et al.[22] used conditional generative adversarial network to generate synthetic samples to balance attack intrusion sample set, solved the problem of data set imbalance caused by insufficient attack samples,and studied the impact of generated attack samples on various machine learning classifiers.
Overall,FDIA attack samples that are constructed from the attack vectors have some common features that are easy for the detection model to find patterns for defense.Also,FDIA attack samples based on artificial intelligence often have strong diversity but large data fluctuations,which are also easy to detect.In order to solve the problem of easy detection of constructed attack samples,we use GAN technology [23] to construct FDIA samples.The proposed scheme can enrich the diversity of FDIA samples,while enhancing their concealment.The generated attack samples can better bypass the defense of detection models,and make the constructed FDIA samples less easily detected.In general,compared to existing works,we make the following novel contributions:
• We design an efficient adversarial network framework to learn the distribution characteristics between normal data and attack data in context of distributed smart grids.Our generated FDIA samples can obtain high concealment effect to effectively bypass the defense of the detection model in public-private network edge.
• Our network structure is designed carefully to allow information to be directly transmitted between network layers, which can effectively solve the problems of gradient vanishing and gradient explosion,and make the network converge faster.Also,different operating parameters are introduced to enrich the diversity of FDIA samples.
• Extensive evaluations are conducted on IEEE 14-bus and IEEE 118-bus systems, and four state-of-the-art detection models are used to validate and analyze the proposed framework.The results indicate that the FDIA data generated by our model outperforms traditional FDIA construction methods in terms of concealment and attack strength,as well as the decrease in FDIA detection accuracy.
The remainder of this paper is organized as follows.Section 2 introduces the related work.The detailed procedure of our proposed model is shown in Section 3.We perform comprehensive experiments to evaluate the performance of proposed model, and present the results and corresponding discussions in Section 4.Finally,Section 5 concludes the paper.
FDIA is a type of network attack targeting data-based systems.In this attack,attackers manipulate or tamper with the data input into the system,with the aim of deceiving,interfering,or disrupting the normal operation of the system.To build invisible and unobservable FDIAs, attackers need to know the complete topology information of the power grid,that is,the Jacobian matrixH∈Rm×nof the power grid topology,the formula is:
whereais the attack vector injected with false data,Zais the measured value after being attacked,andxis the original state estimation value,andeis measurement value error vector.If the attacker designs an attack vectorathat satisfies:
wherec= [c1,c2,...,cn]Tis the non-zero vector that can be designed by the attacker,and the non-zero element indicates that the measurement value of the location is attacked,andis the estimated state value ofZaafter being attacked.Accordingly, the expression of the measured valueZaafter being attacked is
According to the bad data detection mechanism of state estimation,the residual in Eq.(4)is
Obviously, as long as the attacker constructs the attack vectorasatisfying Eq.(2), FDIAs can evade the bad data detection mechanism based on state estimation,thus achieving unobservable FDIA attacks.
With the continuous expansion of the power grid, the amount of data is exploding, and the bottleneck of power public-private edge computing is prominent.Considering that the essence of power public-private edge computing is to decentralize some functions of the power cloud center to the network edge near the sensing data source, it can realize the local/nearby processing of data and related applications.In the context of power big data, edge computing can not only reduce the pressure of data traffic in the cloud center,but also improve the efficiency of data processing.Therefore,the detection of FDIA has gradually begun to be implemented efficiently using edge computing.For instance,facing the harm of FDIA attacks in smart grids,some researchers have attempted to construct efficient detection models to detect FDIA [24].Due to the difficulty in obtaining large-scale FDIA attack samples in CPS, the construction of FDIA attack vectors has become a primary problem to be solved.Some researchers have extensively researched the construction of false data injection attack vectors.Ozay et al.[25]improved the ow sparse attack vector scheme and proposed a distributed attack vector construction scheme.Hao et al.[26]proposed a construction algorithm scheme that can quickly generate highly sparse attack vectors.Bhattacharjee et al.[27]hypothesized that the prior information of the system is unknown, and that attackers can implement stealth attacks based on measurement matrices.They propose a stealth attack vector construction scheme based on independent component analysis to identify mixed matrices.Yu et al.[28]developed a blind FDI attack construction strategy based on principal component analysis(PCA),which utilizes instrument measurements to attack the system with high concealment.
Nevertheless,most of these methods design attack samples through mathematical models,lacking statistical learning for both normal and attack samples.As a result,it is difficult for the constructed attack samples to evade deep neural networks based FDIA detection models.Correspondingly,some FDIA detection methods using machine learning and deep learning technology[29]have been developed sequentially.For example,James et al.[30]introduced discrete wavelet transform and deep neural network to reveal the characteristics of spatial data and obtain the temporal data correlation of FDIA detection.He et al.[31]proposed using a conditional deep belief network to explain the highdimensional temporal behavior characteristics of FDIA with strong concealment, and successfully achieved a 90% accuracy rate for FDIA detection.Ozay et al.[32] proposed semi-supervised and online learning algorithms in machine learning to detect FDIA, which can be used in hierarchical and topological networks in different attack scenarios.In the past two years,Qu et al.[33]designed a scheme based on joint image changes to pre-process changes in measurement data.The dynamic correlation of measurement data between adjacent moments makes the distribution characteristics of measurement data changes more significant,and achieves high detection accuracy.Hu et al.[34]jointly used weighted least squares(WLS)and extended kalman filter(EKF)with an exponential weighting function(WEKF)to detect FDIA,which improved the robustness of state estimation.
Overall,most existing FDIA detection models belong to machine learning-based methods.These methods require a large number of training samples to construct detection models, while the FDIA attack samples in CPS pose significant limitations to the construction of detection models due to their extreme sparsity.This paper attempts to fill this gap by constructing large-scale FDIA attack samples using sparse adversarial learning methods.
The proposed framework is mainly composed of three parts: a data pre-processing module, a training module and a generating data module.The data pre-processing module merge multi-sensor data,including normal data and attack data,where the data samples need to be normalized to adapt to network training.In addition, the training module employs the residual network as conditional generator to generate fake attack data, and further uses full connected network as a conditional discriminator to extract features from attack data and normal data.The generator and discriminator are trained interactively to generate a discriminative model to distinguish attack data.Finally, by repeatedly training the generator and discriminator until convergence, the trained generator in the generating module can be employed to construct new attack data samples.The whole framework of our proposed scheme is shown in Fig.1.
In CPS network system,data may come from different sensors.Since these data are very different in data composition,data accuracy,data transmission delay and frequency refresh,it is necessary to pre-process these multi-source heterogeneous data.
Since each column in the dataset corresponds to different characteristics of the sensor.These different features may have different measurement units and scales.These different scales may cause some features to have a greater impact on data analysis and model training.In order to eliminate the dimensional difference and process each feature more fairly in data analysis and model training,we use the min-max normalization method to linearly map the data to a given range,while maintaining the relative order and distribution shape of the original data.
wherexiis an eigenvalue of the original data,max{x}is the maximum value of the features,and min{x}the maximum value of the features.In general,the min-max normalization method contains two steps:search the minimum min{x}and maximum max{x}of a feature set and then calculate the normalized value by applying the above normalization Eq.(6).
Figure 1:Proposed FDIA attack sample generation framework
In this section,we construct a sparse adversarial training model by combining a generator ResNet and a discriminator FCN.In the generator module, we first input a Gaussian noise matrixZthat conforms to the standard normal distributionPz(z).After learning how to mapPz(z)toPr(z)through the generator network,the noise matrixZgenerates fake attack data.In the discriminator,a few attack data samples and some normal samples are selected as sparse real sample set,which is used to train the model to distinguish the authenticity of fake attack data generated by the generator.The schematic diagrams of ResNet and FCN are shown in Figs.2 and 3,respectively,and their specific structure of both will be presented in the following text.
We firstly use Eq.(7)to build the noise matrixZsatisfying the joint Normal distributionPz(z),and iteratively learn the generated normal distribution matrix vectorZinto ResNet.
whereμ=0 andσ=1 are two parameters of standard normal distribution.Meanwhile,we map the learned distributed feature representation to the sample tag space through a fully connected layer,and then increase the input matrixZfrom 128 dimension to 256 dimension.
Figure 2:Structural diagram of generator
Figure 3:Structural diagram of discriminator
Furthermore,F(xl)of the first residual layer can be obtained by sequentially using fully connected operation FC(·), batch-normalization BN(·) and non-linear transformation with the rectified-linear unit activation function ReLU(·),which is shown in Eq.(8).
The corresponding residual block can be divided into two parts:the direct mapping part and the residual part, wherexl+1is the input for the second residual block,h(xl)is the direct mapping part,andF(xl)is the residual part.
Subsequently,similar to the first residual block,the second residual block also employs the fully connected operation,batch-normalization and ReLU activation function to further process the output featurexl+1.Notably, the fully connected operation in the second residual block needs to perform feature dimensionality reduction.
Finally,passing through a linear layer,the outputG(z)can be re-mapped the same dimensionality as the original dataset.The specific network structure of the Generator ResNet is shown in Fig.4.
Figure 4:The specific structure of the generator and the discriminator
When attack sample is built,a discriminator FCN can be constructed to firstly extract the features from attack data and the normal data,which is considered to distinguish the authenticity of fake attack data.
whereh0is the set of fake attack samples and real data samples output by ResNet,and the symbolain Eq.(15) is a fixed minimum value.Correspondingly, the feature dimension ofh0is firstly adjusted to 256 dimensions through a fully connected hidden layer.Then, the problem of neuron death through Eq.(15)activation function is subsequently solved to enhance the expressiveness of the model.Notably, the dropout layer can make the activation values of certain neurons stop working with a certain probability,enhancing the generalization ability of the model.
Afterwards,h1also passes through the fully connected layer,LeakyReLU layer,and dropout layer again,and generatesh2,whose dimension is correspondingly reduced to 1 for output through a fully connected hidden layer FC(·).The specific network structure of the discriminator is shown in Fig.4.
In the training process,we introduce the WGAN-GP loss function,which provides a more stable training process, reducing issues such as mode collapse.Furthermore, the gradient penalty term in WGAN-GP encourages smooth gradient flow throughout the entire network.This results in better convergence and helps mitigate the problem of vanishing gradients.Therefore,we build the loss function Ltotal(D)by combining the WGAN-GP’s loss[35]L1(D)and WGAN’s discriminator loss L2(D),where the former stands for gradient penalty and the latter represents the discriminator loss.During the training stage,we use neural network to train weightwand use back-propagation to update the parameters until the model convergence.After the fake attack data is generated through the generator,the discriminator can conduct authenticity judgment accordingly, and update the corresponding weight values through loss function calculation.The overall training process is shown in Algorithm 1.
Notably,this process can be recorded by an iteration number.Accordingly,the number of training iterations of the conditional generative adversarial network is 20, the batch size is 500, and the optimizer uses Adam learner.In addition, the default parameters are (0.9, 0.99) operated by the Adam learner.The initial learning rate is 10-4, and the weight of each round will decay.During the training process of the model,the optimal generator model is determined through the optimal objective function.
After each iteration, the generator continuously learns the distribution characteristics of attack data samples and updates the weight values of each hidden layer towards the optimization goal.When the set number of iterations are performed, the generator can exactly generate more realistic attack data.Finally, the input samples meeting standard normal distribution are trained, and the corresponding attack data samples with camouflage and confusion can be outputted by the generator accordingly.
In our experiments,two classical power datasets,the IEEE 14-bus system data and the IEEE 118-bus system data [36,37], are used to evaluate the model performance.The IEEE 14-bus and IEEE 118-bus system data refers to two well-known benchmark datasets that represents the electrical power system of a simplified 14-bus and 118-bus network.They are commonly used in power system analysis and serve as a standard reference for various research studies and algorithm development in the field of power systems.Two datasets provide topology information,bus configurations,and electrical parameters of the power system, including generator data, load data, transmission line parameters,and transformer data.They capture the essential elements and characteristics of a power distribution system, enabling researchers and engineers to analyze and study various aspects of power system operation, stability, and control.The composition information of the IEEE 14-bus system data is shown in Fig.5[36,37],and the IEEE 118-bus system data has similar composition information.
In our experiments,because of the fact that the IEEE 14-bus system data and IEEE 118-bus system data are all data from normal measurement situations,in order to simulate the behavior of attackers and the errors of sensors in reality,we conduct two parts of operations on the IEEE 14-bus and IEEE 118-bus system datasets:attack implementation and measurement noise.
• Attack Implementation:the attack data is usually generated using a false data injection method.The random FDIA is a traditional and classical method, which is characterized by injecting randomly sized false attack data into specific nodes.This attack method has a certain degree of concealment and flexibility.Our scheme integrates the random FDIA attack sample generation method and further integrates the idea of sparse sample adversarial learning to construct a sparse learning based attack sample generation method (SL-GAN).Through the trained generator in the model, we can generate more confusing attack data mixed with normal data characteristics.
• Measurement Noise: given the inevitable dynamic noise in the process of measurement and communication, we also add random Gaussian noise to the measured value and test the performance of different noise derivations.
Given that both the IEEE 14-bus and IEEE 118-bus systems have 15,000 pieces of data, two datasets were constructed based on different numbers of attack data points,including balanced and unbalanced datasets.Experimental comparisons were implemented with different attack intensities.The division ratio between the training set and the testing set is 8:2, including 12000 pieces training samples and 3000 pieces testing samples.
In addition,we employ the precision and recall of the generated outputs as performance evaluation metrics, where precision represents the percentage of error data that we have correctly detected as a percentage of all predicted error data and the recall rate represents the percentage of error data that we have successfully detected as a percentage of all error data.The precision and recall are defined as:
where true positive (TP) denotes the amount of false data detected correctly, true negative (TN)the amount of normal data detected as normal, false positive (FP) as the amount of normal data incorrectly detected as false,and false negative(FN)the amount of false data incorrectly detected as normal.
Figure 5: IEEE 14-bus system.The missing node is the generator node.Due to high intensity protection, their data is usually not easily tampered with, so we do not consider the data of these nodes in our testing
Moreover, we also strike the accuracy andF1score of the generated outputs as performance evaluation metrics.and these two metrics can be correspondingly expressed as:
where the accuracy reflects the proportion of correct classification; the recall rate represents the percentage of error data that we have successfully detected among all error data,andF1is the harmonic mean of recall and accuracy.
In the following experiment, in order to verify that the attack data generated by our model can effectively bypass the detection of the detection model,we conduct a series of comparative experiments.All simulations are performed over a machine with an Intel(R)Core(TM)i5-8300H CPU,a NVIDIA GTX 1050, and 16 GB RAM.In general, when our proposed SL-GAN model is used to generate FDIA attack samples,we set the number of training iterations to 20,and the batch size to 500.The loss function uses WGAN loss and the optimizer is set to Adam learner.The default parameters are(0.9,0.99),and Adam learner performs optimization.The initial learning rate was 10-4and the weights are weakened in each round.Generally,an attack sample for IEEE 14-bus system needs to take 20–30 s,while requiring 600–700 s for IEEE 118-bus system.
Regarding the benchmark method for verifying effectiveness, we compared the attack sample generated by the traditional random FDIA attack generation and our proposed model,respectively.Different attack sample generation methods are tested by several state-of-the-art detection models for experimental comparison.These detection models include SGAT[38],RF[13],SVM[11]and CNN[12], where self-supervised graph attention network(SGAT) introduces a graph attention module to improve the accuracy of FDIA detection.RF is the ensemble learning algorithm based on decision tree, which is easy to implement and has low computational overhead, but it shows very amazing performance in classification and regression.CNN can build a binary detection model using diverse convolutional neural networks.SVM is a maximum margin classifier that constructs a hyperplane(s)in a high-dimensional space.It is widely used since it achieved top performance in some classification problems
Fig.6 presents the test results.From this figure, it can be seen that the loss value of the attack data generated by our proposed mode during the training process,e.g.,Fig.6(b),is higher than that of the attack data generated by the original FDIA, e.g., Fig.6(a), regardless of the initial point of loss value and the stable point.This means that the generated samples by our proposed model are more aggressive.In addition,it can be observed in Fig.6(a),the loss curve converges after 8 iterations,while Fig.6(b)converges more slowly,indicating that our attack data has better camouflage,making it more difficult for the detection model to distinguish.This phenomenon can be easily explained as follows.During the process of training the model,we add a certain proportion of sparse normal sample data,which allows the model to learn a certain distribution of normal data and extract some features of normal samples during training.This results in the generated attack data being accompanied by some normal data features,which is more confusing.Accordingly,the detection model requires more iterations to detect and is more time-consuming.
In addition,we conducted a series of experiments on the IEEE 14-bus and IEEE 118-bus system datasets to verify the effectiveness of the model and its applicability to different datasets.In the testing,we used the four basic detection methods mentioned above:SGAT model,RF model,CNN model,SVM model.We compared and detect the two FDIA attack injection datasets mentioned in Section 4.1 by placing them into different detection models.The corresponding results were shown in Fig.7.We can observe from this figure that the attack data generated using the FDIA based on SL-GAN model is more likely to deceive the four detection models than the data injected directly into the random FDIA attack,resulting in an effective decrease in the accuracy of the detection models.This indicates that the attack data generated by our model has stronger confusion and camouflage.To be specific,for IEEE 14-bus system dataset, the accuracy of the SGAT model has significantly decreased by about 12%,15%approximately decrease for deep learning CNN model,and about 5%and 6%for two machine learning models,RF model and SVM model,respectively.Similarly,for IEEE 118-bus system dataset,our model generated attack data with a reduction of approximately 15%in accuracy for SGAT model,12% for CNN model, 7% for RF and and 5% for SVM model.This indicates that the attack data generated by our model can effectively bypass the detection of the detection model.This is mainly because our model can learn the features of normal data during training stage,the generated attack data thus contains the characteristics of normal data, making it easy to find the difference between attack features and normal features.This greatly confuses the extraction and resolution ability of the detection model,ultimately leading to a decrease in the detection accuracy of FDIA.In addition,our model eliminates data redundancy through ascending and descending dimensions,achieving certain results at IEEE 14-bus and IEEE 118-bus system data,greatly enhancing the applicability of different datasets.
Figure 6:Loss curve changes detected using SGAT model for different attack injection data.(a)Training process using attack data generated by random FDIA model.(b) Training process using attack data generated by SLGAN FDIA model
Figure 7:Comparison of accuracy performance of four detection models after injecting different FDIA attacks in(a)IEEE 14-bus system data and(b)IEEE 118-bus system data
Since our model needs to face attack scenarios of new power systems, e.g., VPP, a large-scale attack samples in this scenario are yet difficult to obtain.To verify the universality of the model,we adjusted the number of data injected into the attack from 270 to 7500, and conduct experimental comparisons on four detection models.The corresponding testing results were shown in Table 1.F1score and accuracy are used as evaluation metrics.We can observe in the table that under the balanced samples,F1scores decreased by 17.7%, 5.4%, 14.4%, and 5.9% for SGAT, RF, CNN, and SVM models, respectively.while the accuracy decreased by 11.0%, 4.7%, 14.7%, and 5.6% in each model, respectively.For unbalanced samples,F1scores decreased by 7.6%, 0.2%, 1.5%, and 4.5% in each model,respectively,and the accuracy decreased by 0.8%,0.8%,2.1%,and 4.1%in each model.Apparently, the attack data generated by our model can achieve more effective attack performance under balanced datasets,but the performance is only slight superior on imbalanced datasets.In fact,this phenomenon can be easily explained.Due to the small proportion of attack samples in imbalanced datasets,and the high cost of identifying samples as attack samples in detection models,resulting in a relatively small number of attack data errors.Similarly, we can also see a good reduction effect in Table 2 of IEEE 118-bus system data experiment.Overall, whether on balanced or unbalanced datasets,the attack data generated by our proposed model can effectively bypass the four state-of-theart attack detection models currently present.
Table 2:Performance comparison of different detection models in IEEE 118-bus system on balanced and unbalanced datasets
In order to further construct attack data with more camouflage and diversity, we carefully investigated the performance of the proposed scheme with different parameters,such as the number of loop iterations of the generator and discriminator,and different ratios of attack samples.Similar to previous experiments,we used attack datasets generated under different parameters for detection,and used four machine learning-based detection models,SGAT,RF,CNN,SVM.The impact of different parameters on the quality of the generated attack data were further discussed in this experiment.
We firstly conducted experimental comparisons on the balanced and unbalanced datasets of IEEE 14-bus system by changing the number of loop iterations of the generator and discriminator.We set the number of iterations to 20, 50, and 80, respectively (e.g., epochs = 20, 50, 80).Similar to previous experiments, we compared the accuracy andF1score to evaluate the performance and different parameters of the proposed attack sample generative model.The corresponding detection results were shown in Fig.8.From the figure,it can be seen that after modifying the parameters,as the number of iterations increases,the attack data generated by our SL-GAN model effectively reduces the accuracy of most detection models.It is easy to explain that the training process of our model aims to minimize the loss function by optimizing the model parameters.As the number of iterations increases,the model has more opportunities to learn and adjust parameters, while also better exploring and learning abstract features and patterns in the data.This enables the model to extract richer and more useful feature representations from both normal and attack data, generating higher quality attack data to deceive the detection model.In addition, we can also observe that not all detection models can have a decrease in detection rate as the number of iterations increases.For instance, our model may reach a saturation point at a certain stage,while further increasing the number of iterations will not significantly improve performance or may lead to overfitting.In general, we can generate highquality attack data by increasing the number of training iterations of our proposed model.These attack samples is more difficult to detect for the detection model.However,the time complexity also increases with the iteration rounds.This conclusion can be experimentally verified in Section 4.4.
Figure 8: (Continued)
In addition,we further discussed the attack intensity,which is defined by the proportion of attack samples in all samples.Obviously,attack intensity is closely related to the ratio of input normal data to attack data, for example, attack intensity 5% represents 5% of attack data and 95% of normal data in the whole samples.We conducted experimental comparisons on the SGAT model [38] using attack data generated with different attack intensities, and use the evaluation metrics mentioned in Section 4.1.The corresponding results were shown in Tables 3 and 4.It can be observed that as the attack intensity decreases,four evaluation metrics show varying degrees of decline.This result is not difficult to understand.Due to the close correlation between attack intensity and the proportion of input normal data and attack data,when the proportion of input attack data decreases,the generated attack data only carry fewer characteristics of attack samples and have more characteristics of normal samples.This can naturally enhance the disguise and concealment of the attack data and further deceive the detection model.
To gain more insight,we further tested the computational complexity of our model when using different parameters.Specifically,we compared the computational complexity of the running time in different situations by changing the iteration number epoch parameter and attack intensity parameter.We can see from the following figure that the lower the attack intensity parameter, the higher the iteration number parameter,and the more time it takes to generate attack data.This can be a natural explanation.The lower the attack intensity,the more normal sample data that participates in training,and the more total samples trained, which naturally increases the time consumption.On the other hand, the more iterations, the more times it takes to find the optimal objective function, and the longer it takes to run.Compared to Fig.9, Table 3 and 4, it can be easily observed that when the attack intensity decreases,the camouflage and concealment of the generated attack data will improve.Increasing the same number of iterations can generate the attack samples with higher quality,but,it is more time-consuming.
Table 3:Performance comparison of attack data generated with different attack intensities on SGAT model under IEEE 14-bus system dataset
Table 4:Performance comparison of attack data generated with different attack intensities on SGAT model under IEEE 118-bus system dataset
Figure 9:Complexity testing of different parameters
Due to the dynamic interaction between information flow and energy flow on both sides of the public and private sectors in virtual power plants, information risk and power risk can overlap and cross propagate on both sides of the public and private sectors.However, some detection models may not be able to fully utilize their detection capabilities when facing strong deceptive attack data designed by attackers.Therefore,this paper proposed a model for generating FDIA attack data based on SLGAN.The model first used time alignment technology to merge some normal sample data with the input attack data,and constructed a generator and discriminator for training.The generator was used to generate false attack data,and the discriminator was used to extract the characteristics of the attack samples and normal samples,and distinguish between false attack data and real data.Finally,the input of the trained generator meets the normal distribution, and generates more camouflage and covert attack data samples.A series of experimental comparisons utilizing four state-of-theart detection models on the load datasets of the IEEE 14-node and IEEE 118-node systems were conducted, validating the attack data samples generated by the proposed model.The experimental results show that compared with the traditional random FDIA attack methods,the proposed attack data generative model can effectively bypass the detection of the detection model so that the detection accuracy is effectively reduced.In the public-private network edge scenarios, this method generates more covert and camouflage attack sample data, which expands the attack sample library and can be effectively used to simulate the methods of possible attack and unknown attack.This lays the foundation for further research on improving the anti-interference ability and detectability of detection models,and ultimately enhances the defense against FDIA attacks in smart grids.
Although our proposed method has shown excellent performance in FDIA detection testing,we should note that our attack data generation scheme mainly uses the load dataset of the IEEE 14-bus and IEEE 118-bus systems to simulate multi-source sensing data on the electrical side.In fact,for the special business scenario of public private interaction virtual power plant,it may be more practical to consider the public and private multi-source data sets.In addition,the time cost of generating attack data samples in our proposed model is relatively high.Producing high-quality attack data samples necessitates a significant time investment to achieve a decrease in detection rates.Future work will aim to address these issues,with a focus on enhancing the quality and efficiency of attack data generation.
Acknowledgement:The authors would like to thank the suggestions for data testing from Dr.Jung Yoon Kim and also thank anonymous reviewers for their valuable suggestions which helped to improve this article.
Funding Statement:This work was supported in part by the the Natural Science Foundation of Shanghai (20ZR1421600), and Research Fund of Guangxi Key Lab of Multi-Source Information Mining&Security(MIMS21-M-02).
Author Contributions:The authors confirm contribution to the paper as follows:study conception and design:Fengyong Li;data collection:Weicheng Shen;analysis and interpretation of results:Zhongqin Bi,Xiangjing Su;draft manuscript preparation:Fengyong Li,Weicheng Shen.All authors reviewed the results and approved the final version of the manuscript.
Availability of Data and Materials:All data used or analyzed during this study are included in this article and its references.
Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.
Computer Modeling In Engineering&Sciences2024年5期