SYSTEM AND METHOD FOR ANALYSING RAILWAY RELATED DATA
20230073361 · 2023-03-09
Inventors
Cpc classification
B61L27/57
PERFORMING OPERATIONS; TRANSPORTING
B61L27/40
PERFORMING OPERATIONS; TRANSPORTING
B61L23/048
PERFORMING OPERATIONS; TRANSPORTING
G06N3/082
PHYSICS
B61L27/53
PERFORMING OPERATIONS; TRANSPORTING
B61L25/04
PERFORMING OPERATIONS; TRANSPORTING
International classification
B61L27/57
PERFORMING OPERATIONS; TRANSPORTING
B61L25/04
PERFORMING OPERATIONS; TRANSPORTING
B61L27/40
PERFORMING OPERATIONS; TRANSPORTING
Abstract
The present invention relates to a method and system using multiple data sources for unsupervised and/or semi supervised algorithms to derive features such as speed of the train, length of the train, type of wagons, etc. Thus, classifying train categories. The invention provides a method and a system configured for analysing railway related vibration data. The invention is configured for collecting at least a first dataset from a sensor applied to the railway infrastructure. Further, it is configured for collecting at least a second dataset from a scheduling component. The at least one subset of the first dataset is curated with the second dataset to obtain first training database. The invention further discloses a method comprising the step of predicting at least a likelihood of one train belonging to at least one train-type.
Claims
1. A method for analysing railway related vibration data, the method comprising the steps of: collecting at least a first dataset from a sensor applied to railway infrastructure, collecting at least a second dataset from a scheduling component, curating at least one subset of the first dataset with the second dataset to obtain a first training database, predicting at least a likelihood of one train belonging to at least one train-type.
2. The method according to claim 1 further comprising the step of connecting the at least one sensor to at least one server, wherein the server comprises at least one processing component.
3. The method according to claim 1 wherein the processing component comprises a memory component configured to store at least one of at least the first dataset and the at least second dataset.
4. (canceled)
5. The method according to claim 1 comprising the step of pre-processing the first dataset, in the processing component.
6. The method according to claim 1 comprising the step of automatically converting the at least one first dataset to at least one time-frequency spectrogram.
7. The method according to claim 1 further comprising the step of unsupervised encoding of the at least one spectrogram to at least one feature map.
8. The method according to claim 1 comprising the step of facilitating the processing component with a neural network (NN) component, wherein the NN component is configured to automatically learn at least one lower-dimensional feature map.
9. The method according to claim 1 further comprising the step of teaching the NN component the at least one lower-dimensional feature map.
10. The method according to claim 1 wherein the method comprises the step of automatically calculating at least one nearest sample neighbour in the lower-dimensional feature map.
11. The method according to claim 1 further comprising the step of using the at least one feature map and the second dataset to label the at least one subset of the first dataset.
12. The method according to claim 1 further comprising the step of iteratively extending the label from the at least one subset of the first dataset to the at least one nearest sample neighbour.
13. The method according to claim 1 further comprising the step of predicting a likelihood of a train being of a certain type using the lower dimensional feature map.
14. The method according to claim 1 further comprising the step of predicting a likelihood of a train being of a certain type using the first training database.
15. A train classification system, the system comprising: a sensor configured to provide at least a first dataset and configured to railway infrastructure, a scheduling component configured to provide at least a second dataset, a server configured to curate at least one subset of the first dataset with the second dataset to obtain a first training database, a processing component configured to classify at least one train type, wherein, the system is configured to execute the method according to any of the method claims.
16. The method according to claim 1 comprising the step of further associating at least one weight with at least one distinctive feature of the train.
17. The method according to claim 1 comprising generating the first training database.
18. The method according to claim 5 wherein the step of pre-processing further comprising at least one of the steps: flagging at least one noisy component of the first dataset, removing at least one exponential wakeup, cutting off the edge of the at least one acceleration trace, stretching the at least one first dataset to a pre-determined size, representing the at least one first dataset as a time-frequency spectrogram.
19. The method according to claim 1 comprising the step of scaling the at least spectrogram value within a pre-determined region.
20. The method according to claim 19 comprising the step of generating at least one spectrogram value using hyperparameter optimization on at least one pre-determined truth dataset.
21. The system according to claim 15 wherein the first dataset comprises vibration signal associated with a motion of a rail vehicle, wherein the vibration signal comprises at least one of: at least frequency data; at least displacement data; at least velocity data; at least acceleration data.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0120]
[0121]
[0122]
[0123]
[0124]
[0125]
[0127]
[0129]
[0131]
[0133]
[0134]
[0135]
DETAILED DESCRIPTION OF THE DRAWINGS
[0136] It is noted that not all the drawings carry all the reference signs. Instead, in some of the drawings, some of the reference signs have been omitted for sake of brevity and simplicity of illustration. Embodiments of the present invention will now be described with reference to the accompanying drawings.
[0137]
[0138] A second sensor 9 is also arranged on another sleeper distant from the first sensor 8. Although it seems just a small distance in the present example, those distances can range from the distance to the neighbouring sleeper to one or more kilometres. Other sensors can be used for attachment to the sleepers as well.
[0139] Another kind of sensor 6 can be attached to the mast 4 or any other structure. This could be another sensor, such as an optical, temperature, even acceleration sensor etc. A further kind of sensor 7 can be arranged above the railway as at the beginning or within the tunnel 5. This could be height sensor for determining the height of a train, an optical sensor, a doppler sensor etc. All those sensors mentioned here and before are non-limiting examples.
[0140]
[0141] In S2, the method can comprise a step of processing the sensor data and converting it to a spectrogram. The method can comprise providing a processing unit configured to process the sensor data. Processing sensor data can comprise converting all the acceleration traces to the same length which can be achieved by cutting off the edges. The cutting of the edges can comprise the method flagging or cropping the trace if the RMS value is lower than a pre-determined acceleration value. The process is further described in the later embodiments.
[0142] The processing can also comprise discarding traces. If a trace is longer than a pre-determined time or lower than a pre-determined RMS value it can be discarded. For example, traces longer than 14 seconds can be cargo trains and can be excluded from classification. Further, the method comprises converting the traces to spectrograms. A spectrogram can be a time-frequency representation of an acceleration trace. The spectrogram can split an acceleration trace or a signal into overlapping windows. Further, a power spectrum density (PSD) of the Fourier transform can be calculated. To obtain constant energy per channel Slaney-style Mel scale can be used. The power spectrum density can then me mapped onto the Mel scale. The next step can be to take the logs of the PSDs at each of the Mel frequencies.
[0143] The processing can also comprise constraining the features of the input within a finite region. This is important because the classifier (as described later) can calculate the distance between two points by the Euclidean distance. If one of the features has a broad range of values, the distance will be governed by this particular feature. Therefore, a finite region can be specified so all features can be normalized such that each feature contributes approximately proportionately to the final distance. A global-maxima and/or a global-minima can be calculated from a first dataset.
[0144] The method can further comprise step S3 for extracting features in an unsupervised manner. The method can comprise learning an embedding for the first data set for dimensionality reduction by training the method to ignore noise. This can be used to group, or segment, datasets with shared attributes in order to extrapolate algorithmic relationships. This can identify commonalities in the data and reacts based on the presence or absence of such commonalities in each new piece of data. Each group can be called a cluster.
[0145] In some embodiments the method comprises providing a scheduling database. The scheduling database can comprise at least one time table of a train. In step S6, the processing unit can comprise receiving the schedule data from the scheduling database. The scheduling database can be configured to be updated automatically or semi automatically. In one exemplary embodiment the method can comprise a step, shown in S7 of curating a small subset of unlabelled data in S2 with the schedule data. Data curation can comprise integration of a subset of data collected from the sensor and the schedule data. A subset of unlabelled data collected from the sensor can further be labelled using the information from the schedule data.
[0146] In some embodiments step S4 can comprise labelling the clusters created in S3 automatically using the information from the labelled subsets from S7. Cluster labelling can further comprise examining the features of the labelled data set per cluster to find a labelling that summarizes a class of each clusters and further can distinguish the cluster from each other.
[0147] The method can further comprise S5 of training a neural network to classify train type on the basis of the labelled and/or unlabelled clusters. The training can be done such that when a ‘new’ unlabelled acceleration trace is fed into the classifier it can predict a likelihood for a train being a certain type. The training can be done in a weekly supervised manner, such that the noise in the labels can be adjusted for.
[0148]
[0149] The latent space, feature space, embedding space representation can comprise a compressed representation of multi-dimensional data and the terms can be used interchangeably. The latent space representation can be a representation of variables that are inferred through an algorithm from other variables that are observed directly. The visualization in
[0150]
[0151] The recall or sensitivity can be the ability of the classifier to find all the positive samples for example, the classifier is recognising 4 Train 1 in a trace containing 12 Train 1 and 2 Train 6. Of the 4 identified as Train 1, 3 actually are Train 1, while the remaining 1 is a Train 6. The classifier's precision in this case can be 0.75 while its recall can be 0.25.
[0152] The f1-score can be a weighted harmonic mean of the precision and recall. The support can be the number of occurrences of each class. The class can be type of a train. For example, Train 1 can be one type or class. Train 2 can be another type or class, etc.
[0153]
[0154]
[0155]
[0156] As discussed in
[0157] After the data has been cleaned or pre-processed it can now be stretched to a standard size. In
[0158]
[0159] Number of Mel: 2.sup.6=64
[0160] Fast Fourier transfer window size: 2.sup.10=1024
[0161] Window forward skip: 2.sup.8=256
[0162] Window type: Hann
[0163] The data cleaned or removed at this step can be stored and be used in other embodiments.
[0164]
[0165]
where z can be a vector of the inputs to the output layer, j can be the indices of the output units. A full classifier can comprise at least one of the at least one classifier architecture and at least one fully connected layer, which can result in the classification. The fully connected layers can connect every neuron in one layer to every neuron in another layer.
[0166]
where z can be a vector of the inputs to the output layer.
[0167] The input layer or input matrix can comprise pixels which can also be interpreted as neuron activations. These neurons can be scaled or normalized by a batch normalization layer 40. The normalization can be such that no activation deviates more than a ‘standard deviation’ of the activation strength. It can further allow each layer of a network to learn by itself a little bit more independently of other layers.
[0168]
[0169]
[0170]