Model structure selection apparatus, method, disaggregation system and program
11635454 · 2023-04-25
Assignee
Inventors
Cpc classification
G06N7/01
PHYSICS
G01R19/2506
PHYSICS
International classification
Abstract
Provided an apparatus that receives time series data from a data storage unit storing time series of sample data or feature values calculated from the sample data, computes a measure indicating change and repetition characteristics of the time series data, based on sample value distribution thereof, selects a state model structure to be used for model learning and estimation, from state models including a fully connected state model and a one way direction state model, based on the measure and stores the selected state model in a model storage unit.
Claims
1. A state model structure selection apparatus comprising: a processor; and a memory storing program instructions executable by the processor is configured to execute the program instructions stored in the memory to receive time series data from a data storage unit that stores time series of sample data or feature values calculated from the sample data; compute a measure indicating change and repetition characteristics of the time series data, based on sample value distribution thereof; and select a state model structure to be used for model learning and estimation, from state models including a fully connected state model and a one way direction state model, based on the measure and stores the selected state model in a storage unit that stores the state model selected, wherein the processor is configured to compute: a first probability of number of cycles for each magnitude value in the time series data, by dividing number of cycles of the magnitude value by total number of cycles in the time series data; and a second probability of number of occurrences for each magnitude value in the time series data, by dividing number of occurrences of the magnitude value by a length of the time series data; compute a correlation coefficient between the first probability and the second probability, and select either the fully connected state model or the one way direction state model, based on the correlation coefficient, as the measure.
2. The state model structure selection apparatus according to claim 1, wherein the processor is configured to execute the program instructions stored in the memory to select the fully connected state model when the correlation coefficient is more than a first threshold value, and select the one way direction state model when the correlation coefficient is less than a second threshold value that is less than the first threshold value.
3. The state model structure selection apparatus according to claim 2, wherein, the processor is configured to execute the program instructions stored in the memory to, when the correlation coefficient is between the second threshold value and the first threshold value, check whether or not the second probability is more than a third threshold value, and select the fully connected state model, when the second probability is less than or equal to the third threshold value, and select the one way direction state model when the correlation coefficient is more than a third threshold value.
4. The state model structure selection apparatus according to claim 3, wherein the third threshold value is equal to the second threshold value.
5. The state model structure selection apparatus according to claim 1, wherein the feature value includes a root mean square (RMS) value of a current signal.
6. A computer based method for selecting a state model structure, comprising: receiving time series data from a data storage unit that stores time series of sample data or feature values calculated from the sample data; computing a measure indicating change and repetition characteristics of the time series data, based on sample value distribution thereof received; selecting a state model structure to be used for model learning and estimation, from state models including a fully connected state model and a one way direction state model, based on the measure; and storing the state model selected in a model storage unit, the method further comprising: computing a first probability of number of cycles for each magnitude value in the time series data, by dividing number of cycles of the magnitude value by total number of cycles in the time series data; computing a second probability of number of occurrences for each magnitude value in the time series data, by dividing number of occurrences of the value by a length of the time series data; computing a correlation coefficient between the first probability and the second probability; and selecting either the fully connected state model or the one way direction state model, based on the correlation coefficient, as the measure.
7. The computer based method according to claim 6, comprising selecting the fully connected state model when the correlation coefficient is more than a first threshold value, while selecting the one way direction state model when the correlation coefficient is less than a second threshold value that is less than the first threshold value.
8. The computer based method according to claim 7, comprising: checking, whether or not the second probability is more than a third threshold value, when the correlation coefficient is between the second threshold value and the first threshold value; and selecting the fully connected state model, when the second probability is less than or equal to the third threshold value, while selecting the one way direction state model, when the correlation coefficient is more than a third threshold value.
9. The method according to claim 8, wherein the third threshold value is equal to the second threshold value.
10. The method according to claim 6, wherein the feature value includes a root mean square (RMS) of a current signal.
11. A non-transitory computer-readable medium storing a program causing a computer to execute processing comprising: receiving time series data from a data storage unit that stores time series of sample data or feature values calculated from the sample data; computing a measure indicating change and repetition characteristics of the time series data, based on sample value distribution thereof received; selecting a state model structure to be used for model learning and estimation, from state models including a fully connected state model and a one way direction state model, based on the measure; and storing the state model selected in a model storage unit, the processing further comprising: computing a first probability of number of cycles for each magnitude value in the time series data, by dividing number of cycles of the magnitude value by total number of cycles in the time series data; computing a second probability of number of occurrences for each magnitude value in the time series data, by dividing number of occurrences of the value by a length of the time series data; computing a correlation coefficient between the first probability and the second probability; and selecting either the fully connected state model or the one way direction state model, based on the correlation coefficient, as the measure.
Description
BRIEF DESCRIPTION OF DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
DETAILED DESCRIPTION
(21) In one of example embodiments, a supervised learning approach is used to learn a state model (model parameters) from given waveform data of an appliance. It is noted that model selection means selecting a structure of the state model. In the example embodiments, it is assumed that FHMM model is used.
(22) In one of the example embodiments, there may be provided two types of structures, wherein a first type is a fully connected state model and a second type is a one way direction state model, though not limited thereto. Out of these two types of state models, selection of a particular state model for given waveform data may be performed.
(23) If a fully connected state model is assumed to be applied to all appliances in disaggregation system, then, there would be an increase in complexity of estimation to disaggregate an aggregate waveform of multiple appliances, such as total sum of current consumption thereof, into an individual signal waveform of each appliance. That is, using fully connected state model for all appliances in disaggregation would have such a disadvantage as an increase in complexity for estimating occurrence of a state at a relevant moment in the appliance from the aggregate waveform, such as current consumption waveform.
(24) In one of example embodiments, a processor with a memory storing a program (instructions) therein may be configured to load and execute the program to execute: input processing that receives time series data from a data storage unit that stores time series of sample data or feature values calculated from the sample data; calculation processing that compute a measure indicating change and repetition characteristics of the time series data, based on sample value distribution thereof; and selection processing that selects a state model structure to be used for model learning and estimation, from state models including a fully connected state model and a one way direction state model, based on the measure and stores the selected state model in a storage unit that stores the state model selected. The state model selection out of fully connected state model and one way direction state model may be performed on a per appliance basis. For example, using combination of two types of state models, an appliance A may be assigned fully connected state model and an appliance B may assigned one way direction state model and so on, respectively, based on given waveform of each of appliance A and appliance B.
(25) Assigning one of two state model structures on a per appliance basis can reduce complexity for estimating a waveform of each appliance in disaggregation that disaggregate an aggregate signal waveform of multiple appliances into a signal waveform of an individual appliance.
(26) Waveform data such as current consumption (for example, RMS (root mean square) current) of an appliance includes mainly two types of patterns. Accordingly, it can be categorized into two types. For example, a first type is a step based pattern and the other is a repetition of same sequence. A step based pattern may include a pattern, such as a current consumption of an appliance, that takes a form similar to so called a step function or a staircase function, with horizontal line segments with jumps in-between. For example, a current consumption of an appliance is kept constant for a time period, in an appropriate measurement scale of current, and then changes instantly to jump to another value. In terms of state model structure, this means that a change of state occurs instantly and a state resides in its own state for some time period.
(27) A second type is a sequence based pattern, in which a current consumption is not constant and changes frequently in response to operation and working of an appliance.
Example Embodiment 1
(28)
(29) The data acquisition unit 101 acquires waveform data (or aggregate waveform) of current measured at a main line or a power distribution board (not illustrated in the drawing) to which one or more appliances may be connected. The data acquisition unit 101 may acquire the waveform data (or aggregate waveform), via one or more communication networks, such as LAN (local area network) and/or WAN (wide area network), from a sensor (not illustrated in the drawing) that samples (measures) a signal and is equipped with a communication function, or from an energy management controller that acquires the waveform data (aggregate waveform).
(30) The data preprocessing unit 102 calculates (extracts) features from the waveform data acquired by the data acquisition unit 101 and saves the calculated features in the data storage unit 103. For example, the data preprocessing unit 102 may calculate a root mean square (RMS), as a feature value, from the waveform data (i.e., sampled value of current signal). More specifically, the data preprocessing unit 102 may calculate time series of RMS values from current waveform data (instantaneous current value) sampled by a sensor at a predetermined sampling frequency. Alternatively, the sensor may provide time series data of such as RMS current that the sensor measures (calculates) to the data acquisition unit 101. Alternatively, the data preprocessing unit 102 may use as the feature value a peak value, an average value, a wave height value, etc. of the waveform in a time domain, or, the waveform itself in the time domain. The data preprocessing unit 102 may, if necessary, decimate or interpolate (re-sample) in time domain sample data of the waveform acquired by the data acquisition unit 101. Alternatively, the data preprocessing unit 102 may transform the waveform data using Fourier transform (Fast Fourier Transform (FFT) or Discrete Fourier Transform (DFT), etc.) into a frequency domain and may calculate a feature value based on frequency spectrum components. The sensor may provide time series data of magnitude of power (such as effective power) that the sensor measures to the data acquisition unit 101. In a still further variation, the data acquisition unit 101 may include the sensor to sample (measure) a signal waveform to obtain the time series data of the signal waveform or directly connected to the sensor.
(31) Initially model training and state model learning may be executed to acquire appliance parameters into a probabilistic model that is a state model graph.
(32) The model structure selection unit 104 selects a state model structure (fully connected state model or one way direction state model) from given waveform data stored in the data storage unit 103.
(33) The selection of a state model structure by the model structure selection unit 104 may be executed prior to a model training phase.
(34) When the training model structure of the state model is selected by the model structure selection unit 104, the model learning unit 105 may learn model parameter of the selected structure of the state model and store the trained model(s) for appliance(s) in the model storage unit 106.
(35) The estimation unit 107 retrieves waveform information from the data storage unit 103 and estimates an individual waveform of each of appliances from an aggregate (synthetic) waveform data such as current consumption of a plurality of appliances.
(36) The estimation unit 107 may use the above described EM algorithm, though not limited thereto. In this case, the estimation unit 107 may include a first unit that performs E-step and a second unit that performs M-step. The first unit (process), based on the waveform data (Y.sub.1, . . . , Y.sub.T) stored in the data storage unit 103, may estimate a state probability in each state of each factor (m) (m=1, . . . , M) of FHMM stored in the model storage unit 106, and supply a estimation result to the second unit. The second unit, using the acquired waveform data: Y.sub.1 to Y.sub.T from the data storage unit 103 and the estimation result from the first unit, may perform learning of FHMM, and update the model parameter of FHMM: ϕ={W.sup.(m), π.sup.(m), P.sup.(m), C}, where π.sup.(m) is a initial state probability, P.sup.(m) is a transition probability, C is a covariance matrix, and W.sup.(m) is a characteristic waveform.
(37) The output unit 108 outputs the estimation result (individual waveform of each of appliances) to an output device such as a display device, a storage device, or a network device (not illustrated in the drawing), wherein the network device may be configured to transmit the result to a remote terminal via a network (not illustrated in the drawing).
(38)
(39)
(40) As described above, a step based waveform data (time series data) as schematically illustrated in
(41)
(42) A state transition matrix B with respect to states (values) p.sub.1 to p.sub.K in
(43) State transition probability P (S.sub.t|S.sub.t-1) for an appliance between states can be given as follows:
P(S.sub.t=p.sub.k|S.sub.t-1)=p.sub.k-1)=P(S.sub.t=w|S.sub.t-1=p.sub.K)=1 (7)
P(S.sub.t=p.sub.1|S.sub.t-1=w)=α (8)
P(S.sub.t=w|S.sub.t-1=w)=1-α (9)
(44) Equation (7) indicates that, when a value (operation state) of a state variable S.sub.t-1 at time stamp t-1 is p.sub.k-1, a probability that a value (state of operation) of a state variable S.sub.t at time stamp t transitions to p.sub.k is 1, for k=1 to K, and a probability that, when a value (operation state) of a state variable S.sub.t-1 at time stamp t-1 is p.sub.K, a value (state of operation) of a state variable S.sub.t at the next time stamp t transitions to w is 1.
(45) Equation (8) indicates that the probability that the value (state of operation) of the state variable S.sub.t at time stamp t becomes p.sub.1 is α (0<α<1), when a value (operation state) of the state variable S.sub.t-1 at time stamp t-1 is w (waiting state).
(46) Equation (9) indicates that the probability that a value (state of operation) of the state variable S.sub.t at time stamp t is w (Waiting state) is 1-α, when a value (operation state) of the state variable S.sub.t-1 at time stamp t-1 is w (wait state).
(47) The state variables S.sub.t and S.sub.t-1 in Equations (7)-(9) may be associated with an individual appliance composing a plurality of appliances whose waveforms are aggregated into an aggregate waveform data. That is, S.sub.t and S.sub.t-1 in Equations (7)-(9) of the one way direction state model may be associated with states of a m-th factor (m-th appliance) S.sup.(m).sub.t and S.sup.(m).sub.t-1 out of total M factors (M appliances), at time stamp t and t-1, in model learning process by the model learning unit 105, after the one way direction state model is selected by the model structure selection unit 104.
(48) As described above, the one way direction state model represents the sequence of patterns in the appliance, as illustrated in
(49)
(50)
(51) The following describes procedures for selecting a state model structure based on two types of state model structures with two examples of time series data.
(52) <Step S101>
(53) Input from the data storage unit 103, a waveform time series data x including [x.sub.i], where i=1,2, . . . , n. In this example, though not limited there to, each sample x.sub.i (i=1, . . . , n) may be a magnitude of a root mean square (RMS) current value calculated by the data preprocessing unit 102 from the current waveform data acquired by the data acquisition unit 101. The RMS current value may be obtained from a current sensor. Or, each sample x.sub.i (i=1, . . . , n) may be a magnitude of a feature value other than a RMS that is calculated by the data preprocessing unit 102. Alternatively, each sample xi (i=1, . . . ,n) may be an amplitude (or absolute value of an amplitude) of sample data of the waveform data acquired by the data acquisition unit 101.
(54) In an example of
(55) In
(56) It is noted that assuming that appliances whose current are aggregated into an aggregate waveform are an appliance A and an appliance B, the time series waveform data of
(57) <Step S102>
(58) Calculate a frequency of occurrence of a sampled value data of the time series data x.
(59)
(60) As shown in
(61)
(62) As shown in
(63) <Step S103>
(64) The total number of cycles is counted.
(65) In the case of
(66) In the case of
(67) <Step S104>
(68) A probability of number of cycles for magnitude value: y.sub.j is calculated by dividing an occurrence of number of cycles of y.sub.j by total number of cycles.
(69)
(70) In the case of
P(y.sub.1.sup.c)=4/21=0.190,
P(y.sub.2.sup.c)=3/21=0.143,
P(y.sub.3.sup.c)=7/21=0.333,
P(y.sub.4.sup.c)=2/21=0.095 and
P(y.sub.3.sup.c)=5/21=0.238.
(71) In the case of
P(y.sub.1.sup.c)=6/21=0.285,
P(y.sub.2.sup.c)=10/21=0.476, and
P(y.sub.3.sup.c)=5/21=0.238.
<Step S105>
(72) The number of times y.sub.j in the time series data x is counted.
(73) In the case of
(74) In the case of
(75) <Step S106>
(76) A probability of number of occurrences of y.sub.j, is calculated by dividing Number of occurrences of magnitude value: y.sub.j by length of x.
(77)
(78) In the case of
P(y.sub.1°)=20/105=0.190,
P(y.sub.2°)=15/105=0.143,
P(y.sub.3°)=35/105=0.333,
P(y.sub.4°)=10/105=0.095, and
P(y.sub.5°)=25/105=0.238.
(79) In the case of
P(y.sub.1°)=30/45=0.667,
P(y.sub.2°)=10/45=0.222, and)
P(y.sub.3°)=5/45=0.111.
<Step S107>
(80) A correlation coefficient [R] between the probability of number of cycles and the probability of number of occurrences of y.sub.j is calculated.
(81)
(82)
(83) <Step S108>
(84) In a case where the correlation coefficient is less than 0.5, a type 2 state model structure is selected (Step S113).
(85) <Step S109>
(86) In a case where the correlation coefficient is greater than 0.8, a type 1 state model structure is selected (Step S112).
(87) <Step S110>
(88) In a case where the correlation coefficient is less than or equal to 0.8 and greater than or equal to 0.5, check if the probability of number of occurrences of y.sub.j is greater than 0.5 (Step S111).
(89) If the probability of number of occurrences of y.sub.j is greater than 0.5, the type 2 state model structure is selected (Step S113).
(90) If the probability of number of occurrences of y.sub.j is less than or equal to 0.5, the type 1 state model structure is selected (Step S112).
(91) The type 1 state model is a fully connected state model as illustrated in
(92) The two types of state model structures are selected and assigned to appliances on a per appliance basis.
(93) The selection method selects a state model structure with the concept of occurrence of y.sub.j and the number cycles of y.sub.j. The correlation coefficient is used to find relationship between these two variables.
(94) The correlation coefficient [R] between 0 and 0.5 is a weak correlation between two given variables.
(95) The type 2 state model structure is selected for this kind of relation (weak correlation) between the probability of number of cycles and the probability of number of occurrences of y.sub.j. The reason for this is that repetition of the same sequence in given data will have the same y.sub.j repeating after each time period. The correlation between the two variables being less than 0.5 may be used as a threshold value to select the type 2 state model structure.
(96) Secondly, a moderate correlation [R] between the two variables lies in between 0.5 and 0.8. When the correlation coefficient [R] between the probability of number of cycles and the probability of number of occurrences of y.sub.j lies in this range, the probability of occurrence of y.sub.i is further examined as one extra step (Step S111 in
(97) The probability of occurrence of any y.sub.j being greater than 0.5 will lead to final selection of the type 2 state model structure.
(98) The reason of adopting 0.5 as a threshold for probability of occurrence of y.sub.j is that if a state is occurring more than 50% of time from the total time period, then it is clear that the structure of one way direction will be suited better to the given waveform data. If the probability of occurrence of y.sub.j is greater than 0.5, then, in terms of state model approach, the number of occurrences of one state is more than 50% of the total length (total time period) of the sampled waveform data length.
(99) In order to reduce complexity, by assigning the type 2 state model structure to this kind of waveform data (time series data), it is better to assign a single state, such as w in
(100) The type 1 model structure will not have a probability 0.5 of occurring the same state as many times, because the full connected topology has many paths which can be chosen from one state and each y.sub.j will share approximately the same probability of occurrence of y.sub.j.
Example Embodiment 2
(101)
(102) A current sensor 53 measures a current at a power distribution line from a distribution board 52 to the loader 514, mounter 515 and the unloader 516, wherein the current is a sum of current supplied to the loader 514, the mounter 515 and the unloader 516, for example. The current sensor 53 transmits the measured current waveform (digital signal waveform) to the waveform disaggregation apparatus 100 via the communication device 54. The current sensor 53 may be constituted by a CT (Current Transformer) (for example, a zero-phase-sequence current transformer (ZCT)), a Hall element or the like. The current sensor 53 may sample a current waveform (analog waveform) with an analog-to-digital converter (not shown), converts it into a digital signal waveform, compresses and encodes it with an encoder (not shown), and then supplies the signal to the communication apparatus 54 as W-SUN (Wireless Smart Utility Network) or the like. The current sensor 53 may calculate RMS(root mean square) current from instantaneously current values sampled at predetermined sampling frequency and transmit time series data of RMS current to the waveform disaggregation apparatus 100 via the communication device 54. The communication apparatus 54 may be disposed in a factory (building). We have found that the one way direction state model can well be applied as a state model structure for the mounter 515.
(103) The disaggregation apparatus 100 may be disposed in a factory or may be mounted on a cloud server connected to the communication device 54 via a wide area network such as the Internet.
(104) The disaggregation apparatus 100 may select as the state model structure for the mounter 515, the one way direction state model as illustrated in
Example Embodiment 3
(105) The disaggregation system or apparatus described in the above embodiments may be implemented on a computer system as illustrated in
(106) The disclosure of the aforementioned PTLs 1-3 and NPL 1 is incorporated by reference herein. The particular exemplary embodiments or examples may be modified or adjusted within the scope of the entire disclosure of the present invention, inclusive of claims, based on the fundamental technical concept of the invention. In addition, a variety of combinations or selections of elements disclosed herein may be used within the concept of the claims. That is, the present invention may encompass a wide variety of modifications or corrections that may occur to those skilled in the art in accordance with the entire disclosure of the present invention, inclusive of claims and the technical concept of the present invention.