Distribution fault location using graph neural network with both node and link attributes
11674994 · 2023-06-13
Assignee
Inventors
Cpc classification
H02H7/008
ELECTRICITY
H02H1/0092
ELECTRICITY
H02H3/021
ELECTRICITY
G01R31/52
PHYSICS
Y04S10/52
GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
G01R31/086
PHYSICS
International classification
G01R31/08
PHYSICS
G01R31/52
PHYSICS
H02H7/00
ELECTRICITY
Abstract
Systems and methods performed by a fault detection apparatus for fault detection and localization in distribution feeders having branches and nodes. The method including receive feeder raw data in a feeder of a power system. Process the feeder raw data with given operational electrical characteristics of the feeder to generate a branch attribute dataset for each branch separated by a pair of nodes for all branches. Generate a node attribute dataset for each node for all the nodes in the feeder. Input the branch and node attribute datasets into a trained neural network to determine whether a branch has a fault and a fault location within the branch, to output a classification of the fault and the fault location. Generate an alert signal based upon determining the classified fault and fault location in response to the alert signal to an outage response system.
Claims
1. A method performed by a fault detection apparatus for fault detection and localization in distribution feeders having branches and nodes, comprising: receiving feeder raw data in a distribution feeder of a power distribution system; processing the feeder raw data with predetermined operational electrical characteristics of the distribution feeder to generate a branch attribute dataset for each branch separated by a pair of nodes for all branches, and generate a node attribute dataset for each node for all the nodes in the distribution feeder; inputting the branch attribute datasets and the node attribute datasets into a trained fault detection neural network to determine whether a branch has a fault and a fault location within the branch, to output a classification of the fault and the fault location, and generate an alert signal based upon determining the classified fault and the fault location; and sending the alert signal to an alert control system, upon the alert signal being received, generate an action in response to the alert signal to an outage response system.
2. The method of claim 1, wherein the fault detection apparatus is configured to measure in real-time pre-fault branch regulations and energizations, and time-synchronized node voltages and currents during a fault, and magnitudes and angles, obtained from sensors associated with the distribution feeder.
3. The method of claim 1, wherein the branch attribute dataset for each branch is generated from real-time measured pre-fault regulations and energizations branch raw data, and the node attribute dataset for each node that is generated from real-time measured during-fault voltages and currents node raw data, via the feeder raw data.
4. The method of claim 1, wherein the branch attributes include at least partial of equivalent nodal conductance and susceptance matrices corresponding to nodes separated by the branch.
5. The method of claim 4, wherein equivalent nodal conductance and susceptance matrices for a distribution line is determined according to a series impedance matrix and a shunt admittance matrix for the line.
6. The method of claim 4, wherein equivalent nodal conductance and susceptance matrices for a distribution transformer is determined according to transformer ratios, series impedances and winding connection for the transformer line.
7. The method of claim 4, wherein equivalent nodal conductance and susceptance matrices for a branch combined a voltage regulator with a downstream distribution line is determined according to a set of regulation ratios and winding connection of the regulator and a series impedance matrix and a shunt admittance matrix of the distribution line.
8. The method of claim 4, wherein equivalent nodal conductance and susceptance matrices for a branch combined a switch with a downstream distribution line is determined according to a set of energized statuses for all phases of the switch and a series impedance matrix and a shunt admittance matrix of the distribution line.
9. The method of claim 1, wherein the node attributes include measured phase to phase voltages and zero-sequence voltages that both include magnitude and angle measurements, and measured injection currents that include magnitude and angle measurements, and the branch attributes include at least partial of equivalent nodal conductance and susceptance matrices corresponding to nodes separated by the branch.
10. The method of claim 1, wherein the predetermined operational electrical characteristics of the distribution feeder include: node standard operation data including load demand profiles and phase connections of power loads connected to nodes; standard operational generation profiles and phase connection of distributed generations connected to nodes; and capacitor capacities and phase connections of shunt capacitors connected to nodes.
11. The method of claim 1, wherein the predetermined operational electrical characteristics of the distribution feeder include: branch data including a series impedance matrix and a shunt admittance matrix for a distribution branch; a set of parameters for each transformer, including transformer ratios, series impedances and winding connection; a set of parameters for each voltage regulator, including regulation ratios and winding connection; and a set of phase energized statues for switches.
12. The method of claim 1, wherein the trained fault detection neural network is trained using a set of fault scenarios generated by simulating a set of pre-determined fault conditions on each branch of all the branches of the distribution feeder separately, and wherein the fault condition includes a fault type, a relative fault location along the branch, an impedance at the fault location, a pre-fault load demand level and a pre-fault generation level.
13. The method of claim 12, further including: obtaining a dataset of node attributes, a dataset of branch attributes, and a set of output attributes for each simulated fault scenario, and wherein output attributes include data to identify nodes separated by the branch having a fault, relative distances between the fault location and the nodes of the fault branch, and a set of fault phases of the fault branch.
14. The method of claim 12, wherein the fault type includes a single phase to ground fault, a double phase to ground fault, a phase to phase fault, a triple phase to ground fault, and a phase to phase to phase fault.
15. The method of claim 1, wherein the trained fault detection neural network is a graph neural network, such that the graph neural network includes a series of graph processing layers for aggregating node and branch attributes into node embeddings, and a series of full-connected prediction layers for estimating fault location according to graph node embeddings.
16. The method of claim 15, wherein a first graph processing layer of the series of graph processing layers, sets node embeddings with node attributes, and the successive graph processing layer calculates its hidden node embeddings as an activated sum of combination of weighted node embeddings at a previous layer and weighted sum of neighborhood impacts, wherein neighborhood impacts for each neighbor is calculated as a decayed combination of neighbor embeddings and weighted branch attributes for the branch connected to neighbor node, and wherein a decay factor is calculated as an activated sum of weighted node embeddings, weighted branch attributes, weighted neighbor embeddings and an addition of biases.
17. The method of claim 16, wherein a sum of neighborhood impacts is approximated as expected neighborhood impacts of a fixed number of neighbor samples, and wherein a sampling probability is approximated according to a norm of the combination of neighbor embeddings and weighted branch attributes.
18. The method of claim 16, wherein each predicting layer calculates its output features as an activated sum of weighted inputs from a previous layer with an addition of biases, wherein the inputs of first predicting layer are the calculated node embeddings of last graph processing layer, and wherein the output features of last predicting layer are data relating to the fault location.
19. The method of claim 16, wherein the summed neighborhood embeddings are estimated by sampling a fixed number of neighbors and approximated as expectations of neighborhood embeddings for samples with sampling probability defined according to a norm of each neighborhood embeddings.
20. The method of claim 1, wherein the received feeder raw data includes real-time measured voltage and current raw data from the nodes that is recorded with a physical intelligent electronic device (IED) or physical phasor measurement unit (PMU).
21. The method of claim 1, wherein the received feeder raw data includes real-time measured regulation and energization raw data from the branches that is recorded with a controller for a switch, or a tap changer for a voltage regulator.
22. A fault detection apparatus for fault detection and localization in distribution feeders having branches and nodes, comprising: a computing system having, a transceiver, data storage with instructional modules and circuitry configured for processing, to cause the apparatus to receive, via the transceiver, feeder raw data in a distribution feeder of a power distribution system; process, via the processor, the feeder raw data with predetermined operational electrical characteristics of the distribution feeder data accessed via the data storage, to generate a branch attribute dataset for each branch separated by a pair of nodes for all branches, and generate a node attribute dataset for each node for all the nodes in the distribution feeder; input the branch attribute datasets and the node attribute datasets into a trained fault detection neural network to determine whether a branch has a fault and a fault location within the branch, to output a classification of the fault and the fault location, and generate an alert signal based upon determining the classified fault and the fault location; and send, via the transceiver, the alert signal to an alert control system, upon the alert signal being received, the alert control system generates an action in response to the alert signal to an outage response system to reroute power and restore service to the disconnected power of the distribution feeder with the fault.
23. A non-transitory computer readable medium, having a computer program thereon, wherein the computer program, when executed by a processor of a fault detection apparatus, causes the processor to: receive feeder raw data including real-time measured pre-fault branch regulations and energizations data, and real-time measured during-fault node voltages and currents raw data, in a distribution feeder of a power distribution system; process the feeder raw data with predetermined operational electrical characteristics of the distribution feeder to generate a branch attribute dataset for each branch separated by a pair of nodes for all branches, and generate a node attribute dataset for each node for all the nodes in the distribution feeder; input the branch attribute datasets and the node attribute datasets into a trained fault detection neural network to determine whether a branch has a fault and a fault location within the branch, to output a classification of the fault and the fault location, and generate an alert signal based upon determining the classified fault and the fault location; and reroute power and restore service to the disconnected power of the distribution feeder with the fault, based upon the alert signal being sent to, and received by, an outage response system.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) The presently disclosed embodiments will be further explained with reference to the attached drawings. The drawings shown are not necessarily to scale, with emphasis instead generally being placed upon illustrating the principles of the presently disclosed embodiments.
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
(21) While the above-identified drawings set forth presently disclosed embodiments, other embodiments are also contemplated, as noted in the discussion. This disclosure presents illustrative embodiments by way of representation and not limitation. Numerous other modifications and embodiments can be devised by those skilled in the art which fall within the scope and spirit of the principles of the presently disclosed embodiments.
DETAILED DESCRIPTION
(22) The present disclosure relates to detection and location of short circuit faults in power distribution systems.
(23)
(24) Referring to
(25) Step 130 of
(26) Step 131 of
(27) Step 132 of
(28) Step 136 of
(29) Referring to
(30)
(31) Still referring to
(32)
(33) When there is a short-circuit fault 275 occurring in any distribution line 220 or bus 1-4, we have to find the exaction location of the fault based on real-time measurements that can be sensed from buses 1-4, and a system topology and branch parameters can be provided by a switch controller, transformer/regulator tap changer, and other information sources of the power distribution system 210.
(34) Still referring to
(35) .sup.12, where |⋅|,∠⋅ stands for an absolute value and a phase angle of a complex number, V.sub.p.sup.x and I.sub.p.sup.x denotes phase voltage and phase injection current at phase x of bus p. Values corresponding to unmeasured phases are set to zero. A data sample of measurements from the grounded distribution system can then be represented as X∈
.sup.n×12, where n is the number of buses.
(36) .sup.14, where, V.sub.p.sup.xy denotes phase-to-phase to-phase voltage between phase x and y of bus p. A data sample of measurements from the ungrounded distribution system can then be represented as X∈
.sup.n×14, where n is the number of buses to be considered.
(37) The distribution system has branches with various types, such as distribution line, transformer, breaker or switch, voltage regulator. In order to use a uniform data set representing various types of branches, we use an equivalent nodal conductance matrix G.sup.eqv and an equivalent nodal susceptance matrix B.sup.eqv to represent branch features of graph neural network.
(38) .sup.72, and the branch parameters of the system can be represented as Y∈
.sup.m×72, where m is the number of branches to be considered.
(39) The branches can be categorized into impedance-based branches, and zero-impedance branches. The impedance-based branches include distribution lines, and transformers. The zero-impedance branches include voltage regulators, circuit breakers, and switches.
(40) Still referring to
(41)
I.sub.ps and I.sub.sp are the vectors of phase currents flowing through bus p, and bus s into the branch respectively. V.sub.p and V.sub.s are the vectors of phase voltages at bus p and bus s. Y.sub.pp and Y.sub.ss are the self admittance matrix at bus p and bus s; Y.sub.ps and Y.sub.ps are the mutual admittance matrices between bus p and bus s, and bus s and bus p. The equivalent nodal conductance and susceptance matrix, G.sup.eqv and B.sup.eqv are the real and imaginary parts of an equivalent nodal admittance matrix Y.sup.eqv:
G.sup.eqv+jB.sup.eqv=Y.sup.eqv (2)
And the equivalent nodal admittance matrix Y.sup.eqv is defined as:
(42)
(43) For a three-phase branch, all branch currents and bus voltages are 3 by 1 vectors, and self and mutual admittance matrices are 3 by 3 matrix.
(44) The equivalent nodal conductance and susceptance matrices for a distribution line is determined according to a series impedance matrix and a shunt admittance matrix for the line.
(45)
(46) Specifically,
(47) The equivalent nodal conductance and susceptance matrices for a transformer is determined according to transformer tap ratios, series impedances and winding connection for the transformer.
(48) Still referring to
(49)
(50)
(51)
wherein I.sub.ms and I.sub.sm are the vectors of phase currents flowing through bus m and bus s, and V.sub.s is the vector of phase voltages at bus s. If the voltage amplifying matrices of the voltage regulator are given in terms of line-to-line voltages (i.e. phase-to-phase voltages), the branch currents and nodal voltages are related as:
(52)
CV.sup.LP is a voltage conversion factor matrix that converted phase-to-ground voltages into phase-to-phase voltages,
(53)
CV.sup.PL is a voltage conversion factor matrix that converted phase-to-phase voltages into phase-to-ground voltages,
(54)
(55) The equivalent nodal conductance and susceptance matrices for a branch combined a switch or breaker with a downstream distribution line is determined according to a set of energized statuses for all phases of the switch or breaker and a series impedance matrix and a shunt admittance matrix of the distribution line.
(56)
(57) The switch branch between bus m and bus p is represented by a phase energized status matrix:
(58)
wherein S.sub.mp.sup.x stands for energized status for phase x, x∈{a, b, c}. S.sub.mp.sup.x equals to 1 if energized, otherwise equals to zero. The merged branch current I.sub.ms flowing through bus in and bus s relates to the original branch current I.sub.ps flowing through bus p to bus s as:
I.sub.ms=S.sub.mpI.sub.ps (7)
(59) Still referring to
V.sub.p=S.sub.mpV.sub.m+(I−S.sub.mp)V.sub.s (8)
(60) Therefore, the currents for the combined branch between bus in and bus s, I.sub.ms and I.sub.sm relates the phase voltages, V.sub.m and V.sub.s as:
(61)
wherein, Y.sup.eqv equivalent nodal admittance defined by the energized status matrix for the switch, S.sub.mp and self and mutual admittances for the impedance branch between bus p and bus s, Y.sub.pp and Y.sub.ss, Y.sub.ps and Y.sub.sp, according to:
(62)
(63) Still referring to
(64) The fault locations are modeled as output features of nodes. There are many ways to define node output features representing the fault conditions. The output features can be represented using either real numbers as shown in
(65) .sup.3, and the output features of the system can be represented as Z∈
.sup.n×3, where n is the number of buses to be considered.
(66) In
(67) For example, in
(68)
and the output features for bus s are
(69)
wherein only the output features corresponding to fault phase A have non-zero values.
(70) For a known fault event, given the fault branch, fault location and fault phases, we can determine output features for all busses accordingly. Therefore, a full set of output features, and node features and branch features for the event can be obtained and served as a training sample for learning a relationship between output features and node and branch features.
(71) Still referring to
(72)
wherein ∥⋅∥ is the cardinality of a set, ô.sub.pp.sup.x and ô.sub.s.sup.x are the estimated output features corresponding to faulted phase x of bus p and bus s, respectively.
(73) .sup.6. The elements of first row of output features, (o.sub.p.sup.UP−x, x∈{a, b, c}) may have non-zero values if a fault occurs upstream to the bus on some of phases, and the magnitudes of non-zero values are related to the distance between fault location to the bus. Similarly, the elements, (o.sub.p.sup.DN−x, x∈ {a, b, c}) of second row of output features for bus p have non-zero values if a fault occurs downstream to the bus on some phases, and the values are related to the distance between fault location to the bus. The output features of the system can be represented as Z∈
.sup.n×6, where n is the number of buses to be considered.
(74) Referring to
(75)
since the fault is downstream to bus p. The output features for bus s are
(76)
since the fault is upstream to bus s. Only the output features corresponding to fault phases B and C have non-zero values.
(77) Still referring to
(78)
wherein ∥⋅∥ is the cardinality of a set, ô.sub.pp.sup.DN−x and ô.sub.s.sup.UP−x are the estimated output features corresponding to faulted phase x of bus p and bus s, respectively.
(79) .sup.6n.sup.
.sup.n×6n.sup.
(80) Still referring to
(81)
wherein ceil(⋅) rounds a number up to the next largest integer, Δd is the length of section of branch between bus p and bus s. Meanwhile, bus s is the downstream bus for the fault branch between bus p and bus s. Then, elements of output feature matrix of bus s at all columns of row n.sub.s are set as 1, n.sub.s represents the location of fault section along the branch with respect to bus s, and is determined according the relative distance from fault location to bus p, according to:
(82)
(83) Still referring to
(84)
wherein {circumflex over (n)}.sub.p.sup.x and {circumflex over (n)}.sub.s.sup.x are the indices of rows of estimated output features that have elements with value 1 at the columns corresponding to phase x of bus p and bus s, respectively. The actual fault spot can be approximated using the mid-point of s.sub.p-th section from bus p toward bus s, and then the ratio of distance between fault spot to upstream bus p over length of the branch, α.sub.p is determined as:
(85)
(86) Based on the node features, branch features and output features defined above, we can formulate the fault location task as a multiple non-linear regression problem if fault locations are represented as output features using real numbers, or multiple-class classification problem if locations are represented using binary numbers. More specifically, given a matrix of sample node features X.sup.(s), and a matrix of sample branch features Y.sup.(s), the vector/matrix of sample fault location Z.sup.(s), is obtained by Z.sup.(s)=f(X.sup.(s), Y.sup.(s)), where f is a specific faulty location regression/classification model, s is the index for the sample fault event. The fault location vector Z.sup.(s) defines the fault indications for all buses, in which the terminal buses for fault branches are set with non-zero real/binary values on faulted phases in which the non-zero values are related to the distance between the fault spot and corresponding bus. A fault is correctly located if {tilde over (Z)}.sup.(s)=Z.sup.(s), where Z.sup.(s) indicates the true fault location, and {tilde over (Z)}.sup.(s) is the estimated fault location corresponding to X.sup.(s) and Y.sup.(s).
(87) The present disclosure can include a graph neural network (GNN) that is used to map the relationship between the fault locations with bus features and branch features of the power distribution system. The graph processing layers with combined node and link attributes are used to map system topology, bus measurements and branch parameters into hidden node embeddings, and full-connected dense layers are used to relevant fault locations to hidden node embeddings. As shown in
(88)
(89) For a given distribution system, normal and faulty cases are simulated for each branch in the system to generate the training and test datasets used for training and evaluating the fault location models. The types of faults include single phase to ground fault, double phase to ground fault, phase to phase fault, and three phase to ground fault, and phase-to-phase-to-phase fault. The different fault locations for each branch, different fault resistances for each fault, and different load levels for the system are simulated. The voltage and current phasors are measured during the fault.
(90) Still referring to
(91) The GNN model used is an extended Graph Convolutional Network model. A Graph Convolutional Network (GCN) has proved to be a powerful architecture in aggregating local neighborhood information for individual graph nodes. Low-rank proximities and node features are successfully leveraged in existing GCNs, however, attributes that graph links may carry are commonly ignored, as almost all of these models simplify graph links into binary or scalar values describing node connectedness. In comparison, the extended GCN model used takes both node and link attributes as inputs.
(92)
(93) Suppose an undirected weighted graph G=(V, E) is used to describe a distribution system, where V is the set of nodes, E is the set of links. A neighbor can be described as an ordered pair, containing a neighbor node and the link connecting it to the central node, i.e. (node, link). In order to capture the interactions within a (node, link) neighbor and adequately incorporate link attributes into node hidden representations, the associated neighbor feature is defined using their tensor product, instead of simply adding or concatenating node and link attributes together. Tensor product of two vectors a and b is calculated as ab.sup.T with shape d.sub.a×d.sub.b, and d.sub.a and d.sub.b are the lengths of a and b. The calculated tensor contains all bilinear combinations of the two attributes, and serves as a fully conjoined feature. Formally, for the central node u connected to node v by a link e.sub.u,v, the corresponding neighbor feature is defined as:
f((v, e.sub.u,v)):=f(v).Math.f(e.sub.u,v) (17)
where u and v are nodes in G, e.sub.u,v is a link from node u to node v. .Math. stands for the operation of tensor product. f(⋅) is the feature of a node, a link or a pair, (v, e.sub.u,v) is a neighbor of node u, i,e, a pair of node v and link e.sub.u,v.
(94) Still referring to ((v, e..sub.,v), (w, e..sub.,w)):=
f((v, e..sub.,v)), f((w, e..sub.,w)))
=
f(v), f(w)
.Math.
f(e..sub.,v), f(e..sub.,w)
(18)
,
stands for the operation of inner product.
(95) Based on the neighbor kernel, a kernel of two l-hop neighborhoods with central node u and u′ can be defined as
(96)
by regarding the lower-hop kernel, .sub.N(l−1)(v, v′), as the inner product of the (l−1)-th hidden representations of v and v′. λ∈[0,1] is a decay factor. N(u) is the set of neighbor nodes of u. By recursively applying the neighborhood kernel, a neural architecture for graphs with both node and link attributes, GCN-LASE (i.e. GCN with Link Attributes and Sampling Estimation) can be defined as a graph processing layer as
(97)
λ.sub.u,v.sup.s,l=σ(WG.sub.node.sup.(l)h.sup.(s,l)(u)+WG.sub.link.sup.(l)f.sup.(s)(e.sub.u,v)+WG.sub.neighbor.sup.(l)h.sup.(s,l)(v)+bG.sup.(l)) (20)
h.sup.(s,0)(u)=f.sub.node.sup.(s)(u) (21)
g.sup.(s,l)(v|u)=h.sup.(s,l−1)(v)⊙sigmoid(WA.sub.link.sup.(l)f.sub.link.sup.(s)(e.sub.u,v)) (22)
g.sup.(s,l)(N(u))=Σ.sub.v∈N(u)λ.sub.u,v.sup.(s,l)g.sup.(s,l)(v|u) (23)
h.sup.(s,l)(u)=σ(WA.sub.node.sup.(l)h.sup.(s,l−1)(u)⊕WA.sub.neighbor.sup.(l)g.sup.(s,l)(N(u))+bA.sup.(l)) (24)
where, ⊙ is the operation of element-wise product, and ⊕ is the operation of concatenating input vectors. σ(⋅) is a nonlinear activation function. The action function is a sigmoid function if a fault location regression model is used, and a SoftMax function if a fault classification model is used. h.sup.(s,l)(u) is the hidden representation of node u in layer l. W.sub.node.sup.(l), WG.sub.link.sup.(l), WG.sub.neighbor.sup.(l) and WA.sub.node.sup.(l), WA.sub.link.sup.(l), WA.sub.neighbor.sup.(l) are the weight parameters in the graph neural network.
(98) For each layer l, the above calculation is executed (L.sub.g−l+1) times with different depth/hop for neighborhood.
(99) Still referring to
o.sub.u.sup.(s,0)h.sup.(s,k)(u) (25)
o.sub.u.sup.(s,k)=σ(WP.sub.node.sup.(k)o.sub.u.sup.(s,k−1)+bP.sup.(k)) (26)
(100) Taken a graph neural network with two graph processing layers and one prediction layers as example, the dimensions of weights and biases for the first graph layer WG.sub.node.sup.(1)/WA.sub.node.sup.(1), WG.sub.link.sup.(1)/WA.sub.link.sup.(1), WG.sub.neighbor.sup.(1)/WA.sub.neighbor.sup.(1) are (n.sub.out.sup.(1), m.sub.node), (n.sub.out.sup.(1), m.sub.link), (n.sub.out.sup.(1), m.sub.mode), and the dimension of bG.sup.(1)/bA.sup.(1) is n.sub.out.sup.(1), wherein n.sub.out.sup.(1) is the pre-determined number of node embeddings for first graph processing layer, m.sub.node and m.sub.link are the numbers of node attributes and branch attributes respectively.
(101) Still referring to
(102) The dimensions of weights and biases for the prediction layer WP.sub.node.sup.(1) are (m.sub.out, n.sub.out.sup.(2)), and the dimension of bP.sup.(1) is m.sub.out, and m.sub.out is the number of output features for fault location.
(103) Still referring to
Loss=Σ.sub.s=1.sup.sΣ.sub.uΣ.sub.m=1.sup.m.sup.
and a squared error loss function is used when the fault location regression model is used:
Loss=Σ.sub.s=1.sup.sΣ.sub.uΣ.sub.m=1.sup.m.sup.
Wherein S is the total number of sample fault events, ô.sub.u,m.sup.(s,L.sup.
(104) Still referring to
(105)
where p.sup.(l)(⋅|u) denotes the sampling probabilities in N(u). We then approximate g.sup.(l)(N (u)) through estimating the expectation. As the sampling process is always unbiased, we look for the optimal probabilities that minimize the estimation variance. According to the derivations of importance sampling, the sampling probabilities can be determined to minimize sampling variation as:
(106)
where ∥⋅∥ is the L2-norm of the vector.
(107) Evaluating the sampling probabilities batch-wisely can be rather inefficient. Under the hypothesis that the network parameters do not dramatically vary from batch to batch, a tradeoff can be made between variance and efficiency by controlling the interval of calculating the optimal distribution. That is, the sampling probabilities for all training nodes are calculated every k batches. Although the calculation may be time-consuming, the batch-averaged time cost will be reduced to 1/k .
(108) Still referring to
(109) Algorithm 1 gives a procedure for sampling all nodes needed for each hop of each graph processing layer. Minibatch for node, B contains nodes that we want to generate representations for. N.sup.(l,k) denotes a deterministic function which specifies a random sample of a node's neighborhood with given number, and we index this function by l and k to denote the fact that the random samples are independent across iterations over l and k. Each set B.sup.(l,k) contains the nodes that are needed to compute the representations of nodes at layer l with search depth, k.
(110) TABLE-US-00001 Algorithm 1 Node sampling Algorithm for graph processing layers Input: Graph G(V, E) Number of graph processing layer L.sub.g Minibatch for node, B Neighborhood sampling function, N.sup.(l,k)(u) Output: Set of nodes for generating representation B.sup.(l,k) 1: for l = 1, ..., L.sub.g do 2: B.sup.(l,0) ← B 3: for k = 1, ..., L.sub.g - l do 4: for u ∈ B.sup.(l,k−1) do 5: B.sup.(l,k) ← B.sup.(l,k−1) ∪ N.sup.(l,k)(u) 6: end for 7: end for 8: end for
(111) Algorithm 2 gives a procedure for minibatch forward propagation for each depth of each graph processing layer. S is the set of samples. At each search depth, nodes aggregate information from their local neighbors with weighted by link attributes, and as this process iterates, nodes incrementally gain more and more information from further reaches of the graph.
(112) TABLE-US-00002 Algorithm 2 Forward propogation Algorithm for graph processing layers Input: Graph G (V, E) Number of graph processing layer L.sub.g Minibatch for node, B Neighborhood sampling function, N.sup.(l,k)(u) Set of nodes for generating representation B.sup.(l,k) Output: Representations for nodes h.sup.(s,l)(u), u ∈ B 1: set using Equation (21) 2: for s = 1, ... , S 3: for l = 1, ..., L.sub.g do 4: for k = 1, ... , L.sub.g − l do 5: for u E B.sup.(l,k−1) do 6: generate a given number of neighbors of u, N(u) 6: calcuate λ.sub.u,v.sup.(s,l) using Eqation (20), v ∈ N(u) 7: calculate g.sup.(s,l)(v|u) using Equaton (22), v ∈ N(u) 8: calculate p.sup.(l)(v|u) using Equation (30), v ∈ N(u) 9: calculate g.sup.(s,l)(N(u)) using equation (29) 10: calculate h.sup.(s,l)(u) using equation (29) 11: end for 12: end for 13: end for 14: end for
(113)
(114) The computing device 900 can include a power source 908, a processor 909, a memory 910, a storage device 911, all connected to a bus 950. Further, a high-speed interface 912, a low-speed interface 913, high-speed expansion ports 914 and low speed expansion ports 915, can be connected to the bus 950. Also, a low-speed connection port 916 is in connection with the bus 950. Contemplated are various component configurations that may be mounted on a common motherboard, by non-limiting example, 900, depending upon the specific application. Further still, an input interface 917 can be connected via bus 950 to an external receiver 906 and an output interface 918. A receiver 919 can be connected to an external transmitter 907 and a transmitter 920 via the bus 950. Also connected to the bus 950 can be an external memory 904, external sensors 903, machine(s) 902 and an environment 901. Further, one or more external input/output devices 905 can be connected to the bus 950. A network interface controller (NIC) 921 can be adapted to connect through the bus 950 to a network 922, wherein data or other data, among other things, can be rendered on a third-party display device, third party imaging device, and/or third party printing device outside of the computer device 900.
(115) Contemplated is that the memory 910 can store instructions that are executable by the computer device 900, historical data, and any data that can be utilized by the methods and systems of the present disclosure. The memory 910 can include random access memory (RAM), read only memory (ROM), flash memory, or any other suitable memory systems. The memory 910 can be a volatile memory unit or units, and/or a non-volatile memory unit or units. The memory 910 may also be another form of computer-readable medium, such as a magnetic or optical disk.
(116) Still referring to
(117) The system can be linked through the bus 950 optionally to a display interface or user Interface (HMI) 923 adapted to connect the system to a display device 925 and keyboard 924, wherein the display device 925 can include a computer monitor, camera, television, projector, or mobile device, among others.
(118) Still referring to
(119) The high-speed interface 912 manages bandwidth-intensive operations for the computing device 900, while the low-speed interface 913 manages lower bandwidth-intensive operations. Such allocation of functions is an example only. In some implementations, the high-speed interface 912 can be coupled to the memory 910, a user interface (HMI) 923, and to a keyboard 924 and display 925 (e.g., through a graphics processor or accelerator), and to the high-speed expansion ports 914, which may accept various expansion cards (not shown) via bus 950. In the implementation, the low-speed interface 913 is coupled to the storage device 911 and the low-speed expansion port 915, via bus 950. The low-speed expansion port 915, which may include various communication ports (e.g., USB, Bluetooth, Ethernet, wireless Ethernet) may be coupled to one or more input/output devices 905, and other devices a keyboard 924, a pointing device (not shown), a scanner (not shown), or a networking device such as a switch or router, e.g., through a network adapter.
(120) Still referring to
(121)
(122) For example, as noted above, the DCS 1051 can use computing devices 1046, 1048, to generate a node attribute dataset based on during-fault voltage and current measurements when a request is received, for example, through a web site that transmits the DCS's requests over the Internet to the central computer 1042. In such instances, the requests can be computed and transmitted by executing computer-executable instructions stored in non-transitory computer-readable media (e.g., memory or storage). It is possible that the central computer 1042 can receive node attribute dataset from those computing devices associated with monitoring devices's 1046, 1048, and receive branch attribute dataset from those computing devices associated with branch regulation and energization data via computing devices 1044, 1050 and 1052.
(123) Still referring to
(124) It is contemplated the hardware processor 1054 can include two or more hardware processors depending upon the requirements of the specific application, wherein the processors can be either internal or external. Certainly, other components may be incorporated with method 1000 including output interfaces and transceivers, among other devices.
(125) It is possible the network 1049 can include, by non-limiting example, one or more local area networks (LANs) and/or wide area networks (WANs). Wherein the networking environments can be similar to enterprise-wide computer networks, intranets and the Internet. Contemplated for all the components mentioned that there can be any number of client devices, storage components, and data sources employed within the system 1000. Each may comprise a single device or multiple devices cooperating in a distributed environment. Further, system 1000 can include one or more data source(s) (not shown). The data source(s) can comprise data resources for training neural networks to express fault location regression or classification functions. The data provided by data source(s) may include during-fault voltage and current measurements, pre-fault branch regulation and energization data, and verified fault types and locations for historical short circuit events.
(126) The present disclosure improves the existing technology and technological field, for example the fields of power distribution system management and control using the intelligent controllers controlled based on the fault detection results fed by the DCS. For example, the computing hardware is activating and deactivating the electrical device, such as a feeder breaker based on a command made by the DCS when a fault location is determined. Specifically, that the components of the systems and methods of the present disclosure are meaningfully applied to improve the control of switchable devices using the computing devices associated with the devices, which in turn, improves the power distribution system management. Further, the steps of the systems and methods of the present disclosure are by computing hardware associated with the electrical device.
(127) Features
(128) An aspect can include the node attributes include measured phase to ground voltages that include magnitude and angle measurements, and measured injection currents that include magnitude and angle measurements. Another aspect is the fault detection apparatus is configured to measure in real-time pre-fault branch regulations and energizations, and time-synchronized node voltages and currents during a fault, and magnitudes and angles, either at a start terminal of the primary feeder, at an end terminal of a primary feeder, and at a low voltage side of distribution transformers associated with the distribution feeder, obtained from sensors associated with the distribution feeder
(129) Another aspect is that the node attributes include measured phase to phase voltages and zero-sequence voltages that both include magnitude and angle measurements, and measured injection currents that include magnitude and angle measurements.
(130) Still another aspect is the branch attributes include at least partial of equivalent nodal conductance and susceptance matrices corresponding to nodes separated by the branch. Wherein equivalent nodal conductance and susceptance matrices for a distribution line is determined according to a series impedance matrix and a shunt admittance matrix for the line. Wherein equivalent nodal conductance and susceptance matrices for a distribution transformer is determined according to transformer ratios, series impedances and winding connection for the transformer line. Wherein equivalent nodal conductance and susceptance matrices for a branch combined a voltage regulator with a downstream distribution line is determined according to a set of regulation ratios and winding connections of the regulator and a series impedance matrix and a shunt admittance matrix of the distribution line. Wherein equivalent nodal conductance and susceptance matrices for a branch combined a switch with a downstream distribution line is determined according to a set of energized statuses for all phases of the switch and a series impedance matrix and a shunt admittance matrix of the distribution line.
(131) Another aspect can be the POE characteristics include node data including typical load demand profiles and phase connections of power loads connected to the node, typical generation profiles and phase connection of distribution generations connected to the node, and capacitor capacities and phase connections of shunt capacitors connected to the node.
(132) Still further, an aspect can be the POE characteristics include branch data including a series impedance matrix and a shunt admittance matrix for a distribution branch; a set of parameters for a transformer, including transformer ratios, series impedances and winding connection; a set of parameters for a voltage regulator, including regulation ratios and winding connection; and a set of phase energized statuses for a switch. Another aspect is that the received measured voltage and current raw data from the nodes is recorded with an intelligent electronic device (IED), or a physical phasor measurement unit (PMU). Still another aspect is the received real-time measured regulation and energization raw data from the branches is recorded with a tap changer for a regulator or a controller for a switch.
(133) An aspect is that the fault detection neural network is trained using a set of fault scenarios generating by simulating a set of pre-determined fault conditions on each branch of all the branches of the distribution feeder separately, wherein the fault condition includes a fault type, a relative fault location along the branch, an impedance at fault location, a pre-fault load demand level and a pre-fault generation level. Further composing obtaining a dataset of node attributes, a dataset of branch attributes, and a set of output attributes for each simulated fault scenario. Wherein output attributes include data to identify nodes separated by the branch having a fault, relative distances between fault location and the nodes of the fault branch, and a set of fault phases of the fault branch. Wherein, the fault type includes a single phase to ground fault, a double phase to ground fault, a phase to phase fault, a triple phase to ground fault, and a phase to phase to phase fault.
(134) Another aspect is the fault detection neural network is a graph neural network, wherein the graph neural network includes a series of graph processing layers for aggregating node and branch attributes into hidden node embeddings, and a series of full-connected prediction layers for estimating fault location according to graph hidden node embeddings. Wherein the first graph processing layer sets node embeddings with node attributes, and the successive graph processing layer calculates its node embeddings as an activated sum of combination of weighted node embeddings at previous layer and weighted sum of neighborhood impacts. Wherein neighborhood impacts for each neighbor is calculated as a decayed combination of neighbor embeddings and weighted branch attributes for the branch connected to neighbor node; wherein a decay factor is calculated as an activated sum of weighted node embeddings, weighted branch attributes, weighted neighbor embeddings and an addition of biases.
(135) Still another aspect is that sum of neighborhood impacts is approximated as expected neighborhood impacts of a fixed number of neighbor samples; wherein sampling probability is approximated according to a norm of the combination of neighbor embeddings and weighted branch attributes. Wherein each predicting layer calculates its output features as an activated sum of weighted inputs from previous layer with an addition of biases; wherein the inputs of first predicting layer are the calculated node embeddings of last graph processing layer. Wherein the output features of last predicting layer are data relating to the fault location. Wherein the summed neighborhood embeddings are estimated by sampling a fixed number of neighbors and approximated as expectations of neighborhood embeddings for samples with sampling probability defined according to a norm of each neighborhood embeddings.
(136) Some embodiments of the present disclosure include a GNN model that is an extension of a conventional graph convolutional network (GCN). The GNN of the present disclosure models a more complete set of factors and parameters that may affect the fault behaviors, and then improves the accuracy of fault detection and location, when compared to conventional GCNs. The conventional GCNs are based on conventional Convolutional Neural Networks (CNNs). The CNNs are Deep Learning algorithms which can take in an input image, assign importance (learnable weights and biases) to various aspects/objects in the image and be able to differentiate one from the other. Whereas the conventional Graph Convolutional Networks (GCNs) are of a type of convolutional neural network (CNN), as noted above, that can work directly on graphs and take advantage of their structural information. The conventional GCNs used for detecting faults and fault locations, are specifically for detecting faults at bus locations, due to the conventional GCNs capability in aggregating local neighborhood information for individual graph nodes. These conventional GCNs leverage low-rank proximities and node features of the graph, however, what is ignored are attributes that graph links may carry, all of these conventional GCNs models simplify graph links into binary or scalar values describing node connectedness to identify neighborships and their influence if weighted in the local neighborhoods, at the bus locations.
(137) Definitions
(138) Short circuit fault: can be an electrical circuit that allows a current to travel along an unintended path with no or very low electrical impedance. This results in an excessive current flowing through the circuit. It is an unavoidable fact that distribution systems are subject to various types of short circuit faults along distribution feeders. Permanent faults cause relay actions that open breakers and de-energize the area surrounding the faulted section of the feeder.
(139) Feeder: The electric distribution feeders can by-non-limiting example, have voltage regulator, an in-line transformer, overhead distribution lines and underground cables of various configurations, several unbalanced spot and distributed loads, and shunt capacitor banks. Also, the feeder have three-phase, double-phase, and single-phase laterals.
(140) Event: is considered some action that caused damage to at least a portion of the power grid, resulting in a potential of, a destabilization of or loss of, power in the power distribution network, which causes an interruption of suppling continuous power either immediately or sometime in a near future. Some examples of events may be considered as natural disaster event (weather, earthquake, etc.), an intentional damaging event (terrorist attack, etc.) or an unintentional damaging event (fallen trees, plane crash, train wreck, etc.).
(141) Power disruption: Can be a power outage or power failures in the power distribution network. Examples of some causes of power failures can include faults at power stations, damage to electric transmission lines, substations or other parts of the distribution system, a short circuit, or the overloading of electricity mains. Specifically, a power outage can be a short or long-term state of electric power loss in a given area or section of a power grid, that could affect a single house, building or an entire city, depending on the extent of the damage or cause of the outage.
(142) Power loads: can be an electrical load is an electrical component or portion of a circuit that consumes (active) electric power. This is opposed to a power source, such as a battery or generator, which produces power. In electric power circuits examples of loads are appliances and lights. Loads may be further classified as critical loads and non-critical loads.
(143) Condition information: from devices may include device energized status, device damage/disconnected status, terminal voltages, and power flows. For example, a current condition information received from the devices can be updated condition information for that moment in time the condition information is received or obtained.
(144) Power distribution grid data: Can include a topology of the power distribution grid, locations of loads and sources, typical profiles for loads and generations, along with labeling the one or more loads as the subset of critical loads and the subset of non-critical loads.
(145) Embodiments
(146) The following description provides exemplary embodiments only, and is not intended to limit the scope, applicability, or configuration of the disclosure. Rather, the following description of the exemplary embodiments will provide those skilled in the art with an enabling description for implementing one or more exemplary embodiments. Contemplated are various changes that may be made in the function and arrangement of elements without departing from the spirit and scope of the subject matter disclosed as set forth in the appended claims.
(147) Specific details are given in the following description to provide a thorough understanding of the embodiments. However, understood by one of ordinary skill in the art can be that the embodiments may be practiced without these specific details. For example, systems, processes, and other elements in the subject matter disclosed may be shown as components in block diagram form in order not to obscure the embodiments in unnecessary detail. In other instances, well-known processes, structures, and techniques may be shown without unnecessary detail in order to avoid obscuring the embodiments. Further, like reference numbers and designations in the various drawings indicated like elements. Also, individual embodiments may be described as a process which is depicted as a flowchart, a flow diagram, a data flow diagram, a structure diagram, or a block diagram. Although a flowchart may describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently. In addition, the order of the operations may be re-arranged. A process may be terminated when its operations are completed but may have additional steps not discussed or included in a figure. Furthermore, not all operations in any particularly described process may occur in all embodiments. A process may correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a function, the function's termination can correspond to a return of the function to the calling function or the main function.
(148) Furthermore, embodiments of the subject matter disclosed may be implemented, at least in part, either manually or automatically. Manual or automatic implementations may be executed, or at least assisted, through the use of machines, hardware, software, firmware, middleware, microcode, hardware description languages, or any combination thereof. When implemented in software, firmware, middleware or microcode, the program code or code segments to perform the necessary tasks may be stored in a machine readable medium. A processor(s) may perform the necessary tasks. Various methods or processes outlined herein may be coded as software that is executable on one or more processors that employ any one of a variety of operating systems or platforms. Additionally, such software may be written using any of a number of suitable programming languages and/or programming or scripting tools, and also may be compiled as executable machine language code or intermediate code that is executed on a framework or virtual machine. Typically, the functionality of the program modules may be combined or distributed as desired in various embodiments.
(149) Embodiments of the present disclosure may be embodied as a method, of which an example has been provided. The acts performed as part of the method may be ordered in any suitable way. Accordingly, embodiments may be constructed in which acts are performed in an order different than illustrated, which may include performing some acts concurrently, even though shown as sequential acts in illustrative embodiments. Further, use of ordinal terms such as “first,” “second,” in the claims to modify a claim element does not by itself connote any priority, precedence, or order of one claim element over another or the temporal order in which acts of a method are performed, but are used merely as labels to distinguish one claim element having a certain name from another element having a same name (but for use of the ordinal term) to distinguish the claim elements. Although the present disclosure has been described with reference to certain preferred embodiments, it is to be understood that various other adaptations and modifications can be made within the spirit and scope of the present disclosure. Therefore, it is the aspect of the append claims to cover all such variations and modifications as come within the true spirit and scope of the present disclosure.