Systems and methods of bad data identification and recovery for electric power systems
11480594 · 2022-10-25
Assignee
- Global Energy Interconnection Research Institute Co. Ltd (Beijing, CN)
- State Grid Corporation of China Co. Ltd (Beijing, CN)
- State Grid Jiangsu Electric Power Co., Ltd. (Jiangsu, CN)
- State Grid Shanxi Electric Power Company (Shyandong, CN)
Inventors
- Yingzhong Gu (San Jose, CA, US)
- Guanyu Tian (San Jose, CA, US)
- Chunlei Xu (San Jose, CA, US)
- Haiwei Wu (San Jose, CA, US)
- Zhe Yu (San Jose, CA, US)
- Di Shi (San Jose, CA)
Cpc classification
G06F18/24147
PHYSICS
Y02E40/70
GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
Y02E60/00
GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
G06F17/16
PHYSICS
Y04S10/22
GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
G01R19/2513
PHYSICS
International classification
G06F17/16
PHYSICS
Abstract
Systems and methods for processing measurement data in an electric power system include acquiring the measurement data by a phasor measurement unit (PMU) coupled to a line of the electric power system, and inputting a plurality of the measurement data within a predetermined time window into a K-nearest neighbor (KNN) for identifying bad data among the plurality of the measurement data, wherein when one of the plurality of measurement data contains a bad datum, the machine learning module sends the bad datum to a denoising autoencoder module for correcting the bad datum, wherein the denoising autoencoder module outputs a corrected part corresponding to the bad datum, and when one of the plurality of measurement data contains no bad datum, the machine learning module bypasses the denoising autoencoder module and outputs the one of the plurality of measurement data as an untouched part.
Claims
1. A method for processing measurement data in an electric power system, the method comprising: acquiring the measurement data from the electric power system; and inputting a plurality of the measurement data within a predetermined time window into a machine learning module for identifying bad data among the plurality of the measurement data, the plurality of the measurement data being vector data acquired by a plurality of phasor measurement unit (PMU) coupled to a plurality of lines of the electric power system, wherein when one of the plurality of measurement data contains a bad datum, the machine learning module sends the bad datum to a denoising autoencoder module for correcting the bad datum, wherein the denoising autoencoder outputs a corrected part corresponding to the bad datum; and when one of the plurality of measurement data contains no bad datum, the machine learning module bypasses the denoising autoencoder module and outputs the one of the plurality of measurement data as an untouched part, wherein the denoising autoencoder module includes a magnitude recovery denoising autoencoder and an angle recovery denoising autoencoder, wherein when the bad datum contains only a bad magnitude, the bad datum is only sent to the magnitude recovery denoising autoencoder for the correction; when the bad datum contains only bad angle, the bad datum is only sent to the angle recovery denoising autoencoder for the correction; and when the bad datum contains both bad magnitude and bad angle, the bad datum is sent to both the magnitude recovery denoising autoencoder and the angle recovery denoising autoencoder for the correction.
2. The method of claim 1, wherein the plurality of the measurement data are arranged in a matrix with vector data arranged in columns.
3. The method of claim 1, wherein the predetermined time window slides over time for inputting measurement data at different time.
4. The method of claim 1, wherein the machine learning module includes a K-nearest neighbor (KNN) algorithm.
5. The method of claim 4, wherein the machine learning module identifies a bad datum by weighted majority vote of a predetermined number of nearest data in terms of Euclidean distance.
6. The method of claim 1, wherein the denoising autoencoder module includes a denoising autoencoder with symmetrical layers of neural network that are trained to reproduce input data at an output thereof.
7. The method of claim 1 further comprising combining the untouched part with the corrected part to form a recovered data stream.
8. The method of claim 1, wherein the measurement data received by the machine learning module are always from a predetermined PMU.
9. The method of claim 1, wherein the measurement data received by the machine learning module are from a first PMU at a first time and a second PMU at a second time different from the first time via a data bus.
10. A system for processing measurement data in an electric power system, the system comprising: measurement devices including a plurality of phasor measurement unit (PMU) coupled to lines of the electric power system for measuring state information at the lines; a processor; and a computer-readable storage medium, comprising: software instructions executable on the processor to perform operations, including: acquiring the measurement data from the measurement devices; and inputting a plurality of the measurement data within a predetermined time window into a machine learning module for identifying bad data among the plurality of the measurement data, wherein when one of the plurality of measurement data contains a bad datum, the machine learning module sends the bad datum to a denoising autoencoder module for correcting the bad datum, wherein the denoising autoencoder module outputs a corrected part corresponding to the bad datum; and when one of the plurality of measurement data contains no bad datum, the machine learning module bypasses the denoising autoencoder module and outputs the one of the plurality of measurement data as an untouched part, wherein the denoising autoencoder module includes a magnitude recovery denoising autoencoder and an angle recovery denoising autoencoder, wherein when the bad datum contains only a bad magnitude, the bad datum is only sent to the magnitude recovery denoising autoencoder for the correction; when the bad datum contains only bad angle, the bad datum is only sent to the angle recovery denoising autoencoder for the correction; and when the bad datum contains both bad magnitude and bad angle, the bad datum is sent to both the magnitude recovery denoising autoencoder and the angle recovery denoising autoencoder for the correction.
11. The system of claim 10, wherein the predetermined time window slides over time for inputting measurement data at different time.
12. The system of claim 10, wherein the machine learning module includes a K-nearest neighbor (KNN) algorithm.
13. The system of claim 10, wherein the measurement data received by the machine learning module are always from a predetermined PMU.
14. The system of claim 10, wherein the measurement data received by the machine learning module are from a first PMU at a first time and a second PMU at a second time different from the first time via a data bus.
15. A method for processing measurement data in an electric power system, the method comprising: acquiring the measurement data by a phasor measurement unit (PMU) coupled to a line of the electric power system; and inputting a plurality of the measurement data within a predetermined time window into a K-nearest neighbor (KNN) for identifying bad data among the plurality of the measurement data, wherein when one of the plurality of measurement data contains a bad datum, the machine learning module sends the bad datum to a denoising autoencoder module for correcting the bad datum, wherein the denoising autoencoder outputs a corrected part corresponding to the bad datum; and when one of the plurality of measurement data contains no bad datum, the machine learning module bypasses the denoising autoencoder module and outputs the one of the plurality of measurement data as an untouched part, wherein the denoising autoencoder module includes a magnitude recovery denoising autoencoder and an angle recovery denoising autoencoder, wherein when the bad datum contains only a bad magnitude, the bad datum is only sent to the magnitude recovery denoising autoencoder for the correction; when the bad datum contains only bad angle, the bad datum is only sent to the angle recovery denoising autoencoder for the correction; and when the bad datum contains both bad magnitude and bad angle, the bad datum is sent to both the magnitude recovery denoising autoencoder and the angle recovery denoising autoencoder for the correction.
16. The method of claim 15 further comprising combining the untouched part with the corrected part to form a recovered data stream.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) Various embodiments of the present disclosure can be further explained with reference to the attached drawings, wherein like structures are referred to by like numerals throughout the several views. The drawings shown are not necessarily to scale, with emphasis instead generally being placed upon illustrating the principles of the present disclosure. Therefore, specific structural and functional details disclosed herein are not to be interpreted as limiting, but merely as a representative basis for teaching one skilled in the art to variously employ one or more illustrative embodiments.
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
DETAILED DESCRIPTION
(17) The present disclosure relates to bad data filtering systems and methods for synchrophasor-based state estimation. Various detailed embodiments of the present disclosure, taken in conjunction with the accompanying figures, are disclosed herein; however, it is to be understood that the disclosed embodiments are merely illustrative. In addition, each of the examples given in connection with the various embodiments of the present disclosure is intended to be illustrative, and not restrictive.
(18) Throughout the specification, the following terms take the meanings explicitly associated herein, unless the context clearly dictates otherwise. The phrases “in one embodiment” and “in some embodiments” as used herein do not necessarily refer to the same embodiment(s), though it may. Furthermore, the phrases “in another embodiment” and “in some other embodiments” as used herein do not necessarily refer to a different embodiment, although it may. Thus, as described below, various embodiments may be readily combined, without departing from the scope or spirit of the present disclosure.
(19) In addition, the term “based on” is not exclusive and allows for being based on additional factors not described, unless the context clearly dictates otherwise. In addition, throughout the specification, the meaning of “a,” “an,” and “the” include plural references. The meaning of “in” includes “in” and “on.”
(20) As used herein, the terms “and” and “or” may be used interchangeably to refer to a set of items in both the conjunctive and disjunctive in order to encompass the full description of combinations and alternatives of the items. By way of example, a set of items may be listed with the disjunctive “or”, or with the conjunction “and.” In either case, the set is to be interpreted as meaning each of the items singularly as alternatives, as well as any combination of the listed items.
(21) In present disclosure, a two-stage machine learning based filtering approach is proposed which not only identifies bad synchrophasor measurements but also, in some extent, recovers or corrects them. Working with LSE as a pre-conditioning scheme, the proposed system and method are highly efficient especially when multiple bad data coexist, which is common in practice. In the first stage, patterns of bad data are identified using a K-Nearest Neighbor (KNN) based classifier. Then a neural network with the denoising autoencoder (DA) architecture is proposed to recover the bad data. Both the KNN classifier and the DA can be pre-trained by PMU measurements and therefore the proposed filter is purely data-driven and model-free. The proposed filter can be carried out at the device level without knowledge of the system and its parameters, and multi-thread parallel processing can be achieved to maximize the processing efficiency for real-time applications.
(22) The remainder of this specification is organized as follows. Section I briefly reviews the basics of linear state estimation and its bad data removal process. Section II introduces the proposed methodology and its implementation for parallel processing in an electric power system. Experimental results and case studies are presented in section III.
(23) Section I. Linear State Estimation
(24) Linear state estimation is a fast state estimation method to obtain the real-time states of power systems by leveraging the linear relationship between the PMU measurements and system voltage phasors. The PMUs are usually installed at the terminals of lines, and their measurements include the 3-phase current and voltage phasors in polar coordinates. Transmission systems are usually considered to be three-phase-balanced in the analysis. Hence, positive sequence measurements can be extracted from the 3-phase measurements through the phase to sequence transformation in equation (1), where V.sub.012 denotes the sequence voltage phasor vector that includes zero, positive and negative sequence labeled as 0, 1, and 2 respectively. V.sub.ABC is the three-phase voltage phasor vector of A, B and C phases directly from PMU measurements. Transmission-level LSE is generally implemented upon the positive sequence measurements.
(25)
(26) For a system with N nodes and L lines, in which some nodes and lines are deployed with PMUs so that there are n voltage measurements and l current measurements, the state vector {right arrow over (x)}∈.sup.N×1 includes the voltage phasors of all nodes. The measurement vector {right arrow over (z)}∈
.sup.(n+l)×1 includes the voltage and current phasors of the terminals with PMU installation. The measurement model of PMU data can be derived from Ohm's law as formulated in equation (2), where A∈
.sup.n×N is the relationship matrix between the state vector {right arrow over (x)} and voltage phasor measurement vector {right arrow over (V)}. If the voltage phasor of node j is the i.sup.th component in the measurement vector of voltage phasors, then A.sub.i,j=1; otherwise A.sub.i,j=0, where A.sub.i,j is the element of A on the i.sup.th row and j.sup.th column. Matrix Y.sub.f∈
.sup.l×N is the from-end system admittance matrix used to calculate the current injection at the “from” end of the measured lines. By combining the voltage and current measurements into one formulation, the measurement model of PMU data can be represented by the complex matrix {dot over (H)} in equation (3).
(27)
(28) Although the model in equation (3) is linear, its components are complex numbers. It can be further expanded into a rectangular-coordinate formulation in equation (4). The corresponding measurement model becomes equation (5), where H.sub.real and H.sub.imag are the real and imaginary part, respectively, of the {dot over (H)} matrix. Matrix {dot over (H)} represents the linear model for linear state estimation in rectangular form.
(29)
(30) Based on the formulation in equation (5), it is possible to solve for the states directly. The solution of x is given in equation (6), where the weighted pseudo-inverse of H is calculated using the Moore-Penrose method (E. H. Moore, “On the reciprocal of the general algebraic matrix,” Bull. Am. Math. Soc., vol. 26, pp. 394-395, 1920). Matrix W∈.sup.(n+l)×(n+l) is a diagonal matrix, of which the diagonal components are weights for the corresponding measurements.
{circumflex over (x)}=(H.sup.TW.sup.−1H).sup.−1H.sup.TW.sup.−1z (6)
(31) The bad data identification and removal criteria of LSE is based on the value of normalized measurement residual r.sub.i.sup.N formulated in equation (7), where r.sub.i denotes the measurement residual of the i.sup.th measurement, calculated from (8). Matrix Ω is the covariance matrix of measurement residual calculated from equation (9). At each iteration of bad data identification and removal, if the highest normalized residual is above 3.0, the corresponding measurement is then regarded as an outlier and can be removed. The threshold 3 indicates that the measurement is 3-sigma away from the estimated value assuming the measurement error is normally distributed, which means there is only a 0.3% chance for such a large deviation of the data point from the estimated value to happen. After removal of the bad data, the LSE is performed again with the updated weight matrix and measurement model for the next iteration until the highest normalized residual is less than three.
(32)
(33) Section II. Exemplary Methodology and Implementation of Propose Bad Data Filter
(34)
(35)
(36) In some embodiments, an input of the bad data filter 200 is a 12×T matrix consisting of raw measurement data vectors within the T-step sliding window. The recommended size of T is 0.2-0.5 times of PMU reporting rate for better filtering performance. Each column is a measurement vector, including the three-phase voltage and current phasors in polar coordinates. Equation (10) demonstrates the input data format. The size of the sliding window we choose in this study is 12 to make the input data a rectangular matrix, but this hyperparameter can be tuned according to needs. Typically the longer the window size, the better capability to identify bad data with longer duration. Longer size, nevertheless, sacrifices the ability to capture system dynamics to some extent.
(37)
(38) As shown in
(39) While there are other definitions of distance that can be adopted in the aforemensioned algorithm, the Euclidean distance is exemplarily chosen for the bad data identification module because of its effectiveness and relatively low computational complexity for high dimensional data. Equation (11) shows the formulation of Euclidean distance between two data point x and y of d dimensions, where x.sub.i and y.sub.i are the i.sup.th dimension of each data point.
Dist(x,y)=√{square root over (Σ.sub.i=1.sup.d(x.sub.i−y.sub.i).sup.2)} (11)
(40) As the identification of bad data is dependent on the weighted majority vote of the j nearest labeled data, the weights should be proportional to the inverse of their distance and regularized to the sum of 1 as formulated in equation (12). Variable w.sub.i denotes the weight on the i.sup.th neighbor, and D.sub.i is the distance between the new data and the i.sup.th neighbor.
(41)
(42) An in-sample training error is denoted by the misclassification rate (MR) formulated in equation (13), where FP is false positive, denoting the number of good data identified as bad data, FN is false negative, denoting the number of bad data identified as good data and N.sub.Total is the total number of instances of the training data set. The sum of FP and FN is the total number of misclassified data points. The evaluation metrics of bad data identification on the testing data set are the precision, recall and F1 score formulated in equation (13), where the TP is true positive, denoting the number of correctly identified bad data. The F1 score is a less biased metric compared to the misclassification rate, considering that the number of correctly identified good data would dominate the misclassification rate as the majority of the data are good data. The relation between precision and recall reflects the tendency of over-kill or under-kill of the classification model.
(43)
(44) As shown in
(45)
(46) The encoder 310 can be interpreted as a nonlinear principal component analysis (PCA) process that transforms the corrupted PMU data into a feature space that enables the decoder 320 of the network to reconstruct from Matthias Scholz and Ricardo Vigário, “Nonlinear PCA: a new hierarchical approach”, in Esann, pages 439-444, 2002. The DA is trained with intentionally corrupted PMU data as input and the correct data as the target of the output so that a trained DA is expected to reconstruct the most accurate PMU data from a corrupted one while filtering out the noise and abnormal value in it.
(47) An error evaluation metric of the DA is the root mean square error (RMSE) between the recovered data and target data as formulated in equation (15), where N is the size of the training data set, x and y represents the recovered data and target data, respectively. The error of all elements within the sliding window is considered, and M denotes the number of input measurements and S is the time span of the sliding window.
(48)
(49)
(50) Another critical requirement of online implementation is the efficiency of data processing. PMU reporting rate is usually 25-60 Hz. Hence, the processing capability of the filter is preferred to match the PMU reporting rate to avoid data stacking for many real-time applications.
(51)
(52)
(53) One or more aspects of at least one embodiment may be implemented by representative instructions stored on a machine-readable medium which represents various logic within the processor, which when read by a machine causes the machine to fabricate logic to perform the techniques described herein. Such representations, known as “IP cores” may be stored on a tangible, machine readable medium and supplied to various customers or manufacturing facilities to load into the fabrication machines that make the logic or processor. Of note, various embodiments described herein may, of course, be implemented using any appropriate hardware and/or computing software languages (e.g., C++, Objective-C, Swift, Java, JavaScript, Python, Perl, QT, etc.).
(54) In some embodiments, a particular software module or component may comprise disparate instructions stored in different locations of a memory device, which together implement the described functionality of the module. Indeed, a module or component may comprise a single instruction or many instructions, and may be distributed over several different code segments, among different programs, and across several memory devices. Some embodiments may be practiced in a distributed computing environment where tasks are performed by a remote processing device linked through a communications network. In a distributed computing environment, Software modules or components may be located in local and/or remote memory storage devices. In addition, data being tied or rendered together in a database record may be resident in the same memory device, or across several memory devices, and may be linked together in fields of a record in a database across a network.
(55) Section III. Case Studies
(56) Two case studies are carried out to validate and demonstrate the proposed machine learning-based bad data filtering scheme. The first case is conducted using the IEEE 14-bus system with synthetic bad data. Experimental results of LSE with and without the proposed filter are compared to demonstrate the performance of the proposed approach. The second case investigates performance of the proposed filter in a real-world system with real PMU measurements, demonstrating its benefits for LSE on large-scale systems.
(57) 3.1 Settings a) The PMU reporting rate for the synthetic data is set to 50 Hz and in the real-world case is 25 Hz. b) Algorithms are implemented in MATLAB 2018a under the Microsoft Windows 10 environment. The simulation is done on a computer with Intel® Core™ i5-6400 CPU@3.00 GHz Processor and 8 GB of RAM. c) Similar to the DA evaluation metrics, the estimation error of LSE is the RMSE between the estimated states and the actual states as formulated in equation (16), where x and y are the estimated and actual states. Variable K here denotes the dimension of data and T is the number of steps.
(58)
(59) 3.2 Results on an IEEE 14-Bus System with Synthetic Data
(60) In this subsection, the proposed approach is tested using an IEEE 14-bus system under a set of random generated bad data scenarios which mimics the observed bad data scenarios in practical systems.
(61) 3.2.1 Data Generation
(62) The measurement data containing bad data are generated according to the following assumptions: a) PMUs are placed on all the 20 lines. The voltage phasors of their from-end buses and the current phasors of the lines are measured. b) The bad data are applied to both voltage magnitude and angular measurements. According to the patterns recognized from real-world measurements, random voltage drops and angular shifts are injected into the simulated three-phase voltage data streams independently. The corrupted phase magnitude drops to zero and the shifted angle complies with the Gaussian distribution of N(0,0.5). c) Bad data injected into different PMUs are independent. d) The number of PMUs that contains bad data simultaneously is controlled. We generated 110 scenarios for each number of simultaneously bad PMU, from 1 to 20. The total number of bad data scenarios is 2200. The PMUs with bad data are randomly selected during scenario generation.
(63) 3.2.2 Model Training
(64) The training data set contains 100 scenarios that cover all 20 simultaneously bad PMU numbers. The total number of data points in the training set is 5000, as each scenario contains a 1-second time window that corresponds to 50 data points.
(65) The hyperparameter K of the KNN model is tuned to 3. We noticed that a larger k compromises the in-sample training accuracy due to the unbalance of the training data set. The majority of the training data points are labeled as good data, therefore, the larger the K, the higher likelihood of a bad data point to be classified as good. The hyperparameter of the autoencoder-based data recovery model is the size of its hidden layer. Considering the input and output size are 1212, we set the hidden layer size to be 300 to avoid overfitting and underfitting.
(66) As mentioned in Section II, one of the advantages of the proposed filter is the parallel processing capability. To achieve the parallelism, each PMU needs its own pre-trained filter. The evaluation metric of KNN is the classification accuracy. DA recovery performance is evaluated upon the RMSE value between the target data and recovered data. Table 1 summarizes the training results of all 20 filters. The PMU indexes correspond to the lines being measured.
(67) TABLE-US-00001 TABLE 1 Training result summary Line index KNN DA 1 99.92% 0.0323 2 99.92% 0.0212 3 99.82% 0.0203 4 99.56% 0.0192 5 99.96% 0.0177 6 99.96% 0.0419 7 99.99% 0.0338 8 99.94% 0.0247 9 99.88% 0.0158 10 99.80% 0.0212 11 99.96% 0.0239 12 99.90% 0.0237 13 99.92% 0.0245 14 99.96% 0.0249 15 99.94% 0.0275 16 99.99% 0.0234 17 99.78% 0.0262 18 99.90% 0.0241 19 99.82% 0.0274 20 99.96% 0.0236
(68) 3.2.3 Simulation Results
(69) The pre-trained filters are applied to the rest of the 2000 bad data scenarios for testing.
(70) Table 2 summarizes the classification performance of the filters under different bad data scenarios. Each bad PMU number contains 100 repetitions of randomly generated bad data scenarios. The average precision of all scenarios is above 90%, while the recall is relatively low, indicating that the bad data identified by the proposed filter has high credibility, but it is possible to miss some of the bad data as well. This feature protects the filter from overkilling good measurements and compromising the accuracy of LSE. The missed bad data can still be removed by the bad data identification and removal of LSE.
(71) TABLE-US-00002 TABLE 2 KNN-based bad data identification performance under different numbers of simultaneously bad PMU Bad PMU TP FP FN TN Precision Recall F1 1 33.68 3.69 8.10 3014.53 90.13% 80.61% 85.10% 2 61.66 4.40 14.54 2979.4 93.34% 80.92% 86.69% 3 87.13 6.54 17.55 2948.78 93.02% 83.23% 87.85% 4 130.70 10.10 24.33 2894.87 92.83% 84.31% 88.36% 5 150.54 10.25 29.91 2869.30 93.63% 83.42% 88.23% 6 180.98 15.37 35.09 2828.56 92.17% 83.76% 87.76% 7 202.86 16.50 38.36 2802.28 92.48% 84.10% 88.09% 8 236.90 20.74 45.47 2756.89 91.95% 83.90% 87.74% 9 254.52 18.50 41.06 2745.92 93.22% 86.11% 89.53% 10 276.78 24.15 47.66 2711.41 91.97% 85.31% 88.52% 11 304.53 28.67 49.17 2677.63 91.40% 86.10% 88.67% 12 326.82 25.65 48.33 2659.20 92.72% 87.12% 89.83% 13 338.68 31.90 57.60 2631.82 91.39% 85.46% 88.33% 14 360.12 36.21 57.77 2605.90 90.86% 86.18% 88.46% 15 381.60 37.69 61.13 2579.58 91.01% 86.19% 88.54% 16 399.80 37.65 56.91 2565.64 91.39% 87.54% 89.42% 17 412.35 38.67 61.47 2547.49 91.42% 87.03% 89.17% 18 428.14 42.59 62.39 2526.88 90.95% 87.28% 89.08% 19 449.75 41.41 61.67 2507.17 91.57% 87.94% 89.72% 20 460.87 43.74 59.86 2495.54 91.33% 88.51% 89.90% Overall 5478.41 494.44 878.36 54348.79 91.72% 86.18% 88.87%
(72)
(73) Besides estimation accuracy, another advantage of the proposed MF is its higher time efficiency. LSE bad data identification and removal are performed in an iterative manner, which consumes a long time to solve when the bad data number is large. With the help of the proposed data pre-processing filter, most of the obvious bad data can be removed in one shot, so that the iteration of LSE bad data removal can be reduced, therefore the overall time consumption becomes less.
(74) Table 3 compares the average time consumption of data pre-processing, which is labeled as MF, LSE with data pre-processing, which is labeled as MF-LSE and LSE. The time consumption of LSE increases as the number of bad PMU increases due to more iterations. Then it saturates to approximately 28 ms because LSE cannot remove bad data on critical measurements, therefore, more bad data would not increase the number of iterations further. On the other hand, with the pre-filtered data, the MF-LSE maintains the average time consumption below 18 ms. This result indicates the proposed filter is capable of removing the majority of bad data effectively regardless of bad PMU number and improving the computational efficiency of LSE. Note that the time consumption of the pre-processing is included in the MF-LSE time.
(75) TABLE-US-00003 TABLE 3 Time consumption comparison (ms) Bad PMU MF MF-LSE LSE 1 3.0 16.9 19.5 2 3.2 17.4 19.8 3 3.1 16.9 20.0 4 3.0 16.8 22.9 5 3.0 16.7 24.1 6 3.1 17.1 25.5 7 2.9 16.8 27.6 8 3.0 16.8 29.6 9 3.1 16.9 28.3 10 3.2 17.1 28.4 11 3.0 16.8 28.5 12 3.0 17.0 27.7 13 3.1 17.0 27.8 14 3.2 17.0 27.5 15 2.9 16.8 27.3 16 3.0 16.9 28.4 17 2.9 16.7 27.9 18 2.9 16.8 28.1 19 3.1 17.0 27.1 20 3.1 17.6 28.4
(76)
(77)
(78) 3.2.4 Sensitivity to Loading Conditions
(79) The filtering performance under different loading conditions are tested to show the robustness of the proposed filter against loading uncertainties.
(80) 3.2.5 Robustness Against Topology Inaccuracy
(81) Topology change is another challenge in power system state estimation. Although topology information is usually assumed known from event detectors in related studies, it might be updated not in time or contains minor errors. Hence, the robustness of data-driven PMU data filtering under inaccurate topology is important. In this study, we consider all 20 N−1 loss of line scenarios in the IEEE 14-bus system to exam the robustness of the proposed filter against topology uncertainties.
(82) Table 4 and Table 5 show the average estimation error and standard deviation of LSE and MF-LSE under the 20 loss-of-a-line scenarios. The average estimation error of the proposed MF-LSE method is lower than LSE in almost every scenario, except for when line 1 is tripped. In terms of numerical stability, MF-LSE yields a lower standard deviation of estimation error than LSE in all scenarios. Hence, it can be concluded that with the aid of the proposed machine learning-based filter, the robustness of LSE against topology inaccuracy is enhanced.
(83) TABLE-US-00004 TABLE 4 Estimation error under N−1 scenarios Line index MF-LSE LSE 1 0.1083 0.0439 2 0.0334 0.0481 3 0.05758 0.0715 4 0.0440 0.0711 5 0.0299 0.0709 6 0.0308 0.0713 7 0.03906 0.0714 8 0.0383 0.0785 9 0.0235 0.0712 10 0.0422 0.0717 11 0.0211 0.0720 12 0.0304 0.0738 13 0.0351 0.0725 14 0.0429 0.0789 15 0.0340 0.0773 16 0.0347 0.0735 17 0.0279 0.0818 18 0.0256 0.0737 19 0.0291 0.0715 20 0.0322 0.0715
(84) TABLE-US-00005 TABLE 5 Standard deviation of estimation error under N−1 scenarios Line index MF-LSE LSE 1 0.0073 0.018547 2 0.0064 0.01919 3 0.0045 0.022011 4 0.0059 0.021679 5 0.0054 0.021924 6 0.0053 0.021801 7 0.0068 0.021915 8 0.0041 0.039112 9 0.0039 0.021702 10 0.0063 0.021494 11 0.0041 0.0219 12 0.0045 0.0221 13 0.0039 0.0223 14 0.0034 0.0218 15 0.0042 0.0369 16 0.0039 0.0222 17 0.0025 0.0215 18 0.0041 0.0219 19 0.0050 0.0219 20 0.0040 0.0220
(85) 3.3 Results on a Real-World System with Real PMU Data
(86) An embodiment of the present disclosure has been applied to a practical provincial power grid—the Jiangsu power grid in China. The Jiangsu power grid, as shown in
(87) A triple-circuit transmission line 00JDMB-0DS1X is used as an illustrative example. The three circuits are labeled as line I, II, and III, respectively. The objective of this test case is to identify and recover the bad data on line I using the proposed machine learning-based filter. In order to show the effectiveness of the proposed approach, a following linear state estimator is not included in this example.
(88)
(89) Table 6 summarizes the bad data identification results and the evaluation metrics are listed in Table 7. All identification metrics are above 80%, among which the precision is over 90%, indicating a high identification performance. The RMSE of the recovered data is extremely low, indicating a high recovery fidelity.
(90) TABLE-US-00006 TABLE 6 Classification results summary of line I in the practical case Label Prediction Bad data Good data Bad data 83 6 Good data 16 295
(91) TABLE-US-00007 TABLE 7 Identification accuracy of current angle of line I in the practical case Item Score Precision 93.26% Recall 83.84% F1 88.30% RMSE 0.0045
(92) Table 8 compares the performance of various approaches. MF refers to the proposed machine learning filter (KNN/DA). MF-LSE refers to applying the proposed machine learning filter to a linear state estimator with bad data identification and removal. LSE refers to a linear state estimator with bad data identification and removal with a time limit, beyond which the iteration stops. LSE* refers to linear state estimation with unlimited time to perform bad data identification and removal. The evaluation metrics are listed in the left column. The “Max Residual” is the maximum normalized residual obtained in the final iteration of the LSE. For all approaches, we assume a maximum 200 ms time limit for finishing the computation of each snapshot (except for LSE*), which requires generating an LSE solution five times per second. From Table 8, it is observed that the proposed filter with an LSE gives the highest F1 score, which indicates an overall good performance over precision and recall. The LSE only approach gets a very high precision because it is unlikely to label good data as bad ones. However, due to its iterative structure that an extra round of iteration needed to remove a bad measurement, it can only process a limited number of bad data within 200 ms so that the recall is much lower than using the proposed approach as a data preprocessor. If enough time is given, the linear state estimation will be able to identify most of the bad data and generates a very high F1 score, as is suggested by the LSE* column. Overall, the proposed approach plus an LSE following it gives the highest F1 score in practical systems. If only the proposed approach is used without an LSE following it, the recall score will be compromised.
(93) TABLE-US-00008 TABLE 8 Performance comparison of various approaches Item MF MF-LSE LSE LSE* Precision 0.931 0.9417 1 1 Recall 0.8265 0.9898 0.1719 1 F1 0.8757 0.9652 0.2933 1 Max. Res. 0 2.9422 6.7511 2.9312
(94) Table 9 demonstrates the average computation time for each snapshot given the number of bad data specified in the leftmost column. The row header has the same definition as Table 8. If there is few bad data encountered, the LSE approach gives the fastest performance to obtain a reasonable state estimation solution. As the number of bad data increases, the time consumption for LSE to process all the bad data grows much faster than the proposed method. If only using the proposed data filter without linear state estimation, it gives the best computation speed, but the F1 score will be compromised, as is shown in Table 8. Therefore, a good trade-off between the computation time and solution quality is to use the proposed machine learning filter as a data pre-processor followed by a linear state estimator to calculate the states and process any bad data which is not identified by the proposed approach.
(95) TABLE-US-00009 TABLE 9 Average computation time (ms) for processing bad data Bad Data # MF MF-LSE LSE 0 6.7209 13.6849 6.9615 1 7.1674 14.1444 7.4473 10 7.4171 30.879 79.472 50 12.8356 98.5046 388.1276
(96) Publications cited throughout this document are hereby incorporated by reference in their entirety. While one or more embodiments of the present disclosure have been described, it is understood that these embodiments are illustrative only, and not restrictive, and that many modifications may become apparent to those of ordinary skill in the art, including that various embodiments of the inventive methodologies, the illustrative systems and platforms, and the illustrative devices described herein can be utilized in any combination with each other. Further still, the various steps may be carried out in any desired order (and any desired steps may be added and/or any desired steps may be eliminated).