COMPOUND NEURAL NETWORK ARCHITECTURE FOR STRESS DISTRIBUTION PREDICTION
20210174198 · 2021-06-10
Inventors
Cpc classification
G05B2219/21002
PHYSICS
G06V30/18057
PHYSICS
International classification
Abstract
A neural network architecture and a method for determining a stress of a structure. The neural network architecture includes a first neural network and a second neural network. A neuron of last hidden layer of the first neural network is connected to a neuron of a last hidden layer of the second neural network. A first data set is input into the first neural network. A second data set is input into the second neural network. Data from the last hidden layer of the first neural network is combined with data from the last hidden layer of the second neural network. The stress of the structure is determined from the combined data.
Claims
1. A method of determining a stress of a structure, comprising: inputting a first data set into a first neural network; inputting a second data set into a second neural network; combining data from a last hidden layer of the first neural network with data from a last hidden layer of the second neural network; and determining the stress of the structure from the combined data.
2. The method of claim 1, wherein combining the data further comprises combining data from an i.sup.th neuron of the last hidden layer of the first neural network with data from an i.sup.th neuron of the last hidden layer of the second neural network.
3. The method of claim 1, wherein combining the data further comprises at least one of a scalar mathematical operation and/or a matrix operation.
4. The method of claim 1, further comprising inputting a third data set into a third neural network and combining data from the last hidden layer of the first neural network, data from the last hidden layer of the second neural network and data from the last hidden layer of the third neural network.
5. The method of claim 1, further comprising obtaining the stress for the structure, splitting the stress into a first stress component that is a function of spatial coordinates and a second stress component that is a function of geometry and loading, and inputting the first stress inputs into the first neural network and the second stress inputs into the second neural network.
6. The method of claim 1, wherein the first data set and the second data set are one of intersecting data sets and non-intersecting data sets.
7. The method of claim 1, wherein the structure is loaded by at least one of a mechanical load, a thermal load, and an electromagnetic load.
8. The method of claim 6, further comprising determining at least one of a strain of the structure, a displacement of the structure, a temperature of the structure, a heat flux of the structure, a magnetic flux of the structure, a numerical analysis of the structure, and a measured output of the structure.
9. The method of claim 1, wherein one of the first data set and the second data set is an image or video data set and the other of the first data set and the second data set is a measurement or a numerical data set.
10. The method of claim 1, wherein at least one of the first neural network and the second neural network is at least one of a convolution neural network (CNN), an artificial neural network (ANN), and a recurrent neural network (RNN).
11. The method of claim 1, further comprising determining the stress of the structure from one of a single output and a plurality of outputs.
12. A neural network architecture for determining a stress of a structure, comprising: a first neural network configured to receive a first data set of the structure; a second neural network configured to receive a second data set of the structure; wherein a neuron of last hidden layer of the first neural network is connected to a neuron of a last hidden layer of the second neural network in order to combine data from the respective neurons to determine the stress of the structure.
13. The neural network architecture of claim 12, wherein an i.sup.th neuron of the last hidden layer of the first neural network is connected to an i.sup.th neuron of the last hidden layer of the second neural network.
14. The neural network architecture of claim 12, wherein the neuron of the last hidden layer of the first neural network is connected to the neuron of the last hidden layer of the second neural network to enable at least one of a scalar mathematical operation and/or a matrix operation of the data of the respective neurons.
15. The neural network architecture of claim 12, wherein the first data set of the structure is a function of coordinates and the second data set of the structure is a function of geometry and loading.
16. The neural network architecture of claim 12, wherein the first data set and the second data set are one of intersecting data sets and non-intersecting data sets.
17. The neural network architecture of claim 12, wherein the structure is loaded by at least one of a mechanical load, a thermal load, and an electromagnetic load.
18. The neural network architecture of claim 12, wherein one of the first data set and the second data set is an image or video data set and the other of the first data set and the second data set is a measurement or a numerical data set.
19. The neural network architecture of claim 12, wherein at least one of the first neural network and the second neural network is at least one of a convolution neural network (CNN), an artificial neural network (ANN) and a recurrent neural network (RNN).
20. The neural network architecture of claim 12, wherein the combined data is provided as one of a single output and a plurality of outputs.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0009] Other features, advantages and details appear, by way of example only, in the following detailed description, the detailed description referring to the drawings in which:
[0010]
[0011]
[0012]
[0013]
[0014]
[0015]
[0016]
[0017]
[0018]
[0019]
DETAILED DESCRIPTION
[0020] The following description is merely exemplary in nature and is not intended to limit the present disclosure, its application or uses. It should be understood that throughout the drawings, corresponding reference numerals indicate like or corresponding parts and features.
[0021] In accordance with an exemplary embodiment,
[0022] In operation, the input data for structural component stress prediction can be split into two data sets. In one embodiment, the first set of data includes spatial coordinates and the second set of data includes inputs pertaining to geometric parameters (of the structural part), loads/boundary conditions, and material properties. The first set of data is input to the first neural network 102 and the second set of data is input to the second neural network 104.
[0023] The first neural network 102 is combined with the second neural network 104 via end-to-end neuron addition. In one embodiment, the output from the last hidden layer of the first neural network is combined with the output from the last hidden layer of the second neural network in order to calculate an output parameter, such as a stress on the structural component. It is to be understood that the output in general can be any parameter from FEA solutions such as stress, strain, displacement, temperature, magnetic flux, etc.
[0024] In end-to-end neuron addition, as shown in
[0025] It is to be understood that the neural network architecture 100 can include more than two neural networks with the set of input data being divided amongst the more than two neural networks according to a selected guideline or procedure based on physical laws or other criteria. For more than two neural networks, the end-to-end neural addition includes combining the outputs from first neuron of the last hidden layers of the neural networks, combining the outputs from second neuron of the last hidden layers, etc.
[0026]
z=Σ.sub.iw.sub.ix.sub.i+b Eq. (1)
where w.sub.i is a weight coefficient and b s a bias term. The summation results in a scalar value z. The neuron then activates the scalar value using an activation function G(z), presenting the scalar value z as input to a subsequent neuron. Some illustrative activation functions for use in a neural network are shown in
[0027] For a neural network architecture 100 having k neural networks that are combined via end-to-end neuron addition, the output 130 of the neural network architecture 100 can described as in shown in Eq. (2):
output(y.sub.1x1)=W.sub.nx1.sup.T(Σ.sub.i=1.sup.kh.sub.nx1.sup.i)+b.sub.1x1 Eq. (2)
where h.sub.nx1.sup.i indicates the vector of size n consisting of the last layer of the i.sup.th neural network. W.sup.T.sub.nx1 is a column vector including weight coefficients for each of the summation terms and b.sub.1x1 is the bias term. In vector form, the summation term forms a column vector including n row (one for each neuron), each row including a summation of consisting of k terms (one term for each neural network), as in Eq. (3):
For the illustrative neural network architecture 100 of
[0028] In another embodiment of end-to-end neuron addition, the individual weight coefficients of each neuron are included in the summation term as shown in Eq. (5):
output(y.sub.1x1)=W.sub.pre.sub.
where h.sub.nx1.sup.i indicates the vector of size n consisting of the last layer of the i.sup.th neural network. W.sub.pre.sub.
For only two neural networks (k=2), the column vector of Eq. (6) reduces to
[0029]
[0030] In one embodiment, the new weights and biases 412 can be computed by:
w.sub.new=w.sub.old−αΔw Eq. (8)
and
b.sub.new=b.sub.old−αΔb Eq. (9)
[0031] where α is user-defined learning rate.
[0032]
[0033]
[0034]
NN.sub.1=σ(X)=c=F/A.sub.1 Eq. (10)
The geometry (A) and applied force (F) is input into the second neural network. The output function of the second neural network 104 after learning from training data is given by:
NN.sub.2=σ(A,F)=F(A.sub.1−A)/(A*A.sub.1) Eq. (11)
The end-to-end neural addition of the outputs of NN1 and NN2 gives a total stress as indicated in Eq. (12) for an input area A.sub.1 and applied force (F) is:
σ=NN.sub.1+NN.sub.2=F/A.sub.1 Eq. (12)
[0035] If area A.sub.2 and applied force (F) is inputted to the neural network architecture 100 for the above one-dimensional case, the output of the first neural network 102, which is based on spatial coordinates, remains unchanged, as shown in Eq. (13):
NN.sub.1=σ(X)=F/A.sub.1 Eq. (13)
Meanwhile, the output of the second neural network 104 is given by:
NN.sub.2=σ(A.sub.2,F)=F(A.sub.1−A.sub.2)/(A.sub.1*A.sub.2) Eq. (14)
From end-to-end neural addition, the total determined stress of the deformed object is given as shown in Eq. (15):
σ=NN.sub.1+NN.sub.2=F/A.sub.2 Eq. (15)
[0036]
NN.sub.1=c.sub.1*Sig(a−x+δ)=F/A.sub.1*Sig(a−x+δ) Eq. (16)
where Sig is a sigmoid function based on a length of the first area that limits the stress output to first cross-section 802 and δ is a small number as compared to values of a and b. Meanwhile, the stress function estimated by a second neuron of the first network after learning from training data is given by Eq. (17):
NN.sub.1=c.sub.2*Sig(x−a−δ)=F/A.sub.2*Sig(x−a−δ) Eq. (17)
[0037] The second neural network estimates stress as a function of geometry and load. The output stress function from first neuron of the second neural network 104 after learning from training data is given by Eq. (18):
NN.sub.2=F(A−A.sub.1)/(A*A.sub.1) Eq. (18)
while the output stress function from second neuron of the second neural network 104 after learning from training data is given by Eq. (19):
NN.sub.2=F(A−A.sub.2)/(A*A.sub.2) Eq. (19)
[0038] If input areas of A.sub.3 and A.sub.4 are given as input to the neural network architecture 100 for one dimensional case with two different cross-sections. The resultant stress resulting from end-to-end neuron addition is given by Eq. (20):
σ(x)=F/A.sub.1*Sig(a−x+δ)+F(A.sub.3−A.sub.1)/(A.sub.3*A.sub.1) Eq. (20)
or by Eq. (21):
σ(x)=F/A.sub.2*Sig(x−a−δ)+F(A.sub.4−A.sub.2)/(A.sub.3*A.sub.2) Eq. (21)
where the stress is depending on the x value. Similar logic can be extrapolated to two-dimensional and three-dimensional solid structures.
[0039]
[0040]
[0041] While the neural network architecture has been discussed without specification of the types of neural networks, it is to be understood that either of the first neural network and the second neural network can be a convolution neural network (CNN), a recurrent neural network (RNN), a recurrent neural network (RNN or other suitable neural networks. Additionally, the data is not confined to stress data and can be any selected data set. The stress data can be in different from such as von-misses, maximum principal, etc. In one example, the one set of data can be image/video data while the other set of data is measurement/numeric data.
[0042] While the above disclosure has been described with reference to exemplary embodiments, it will be understood by those skilled in the art that various changes may be made, and equivalents may be substituted for elements thereof without departing from its scope. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the disclosure without departing from the essential scope thereof. Therefore, it is intended that the present disclosure not be limited to the particular embodiments disclosed, but will include all embodiments falling within the scope thereof.