Model-driven deep learning-based seismic super-resolution inversion method
11226423 · 2022-01-18
Assignee
Inventors
- Jinghuai Gao (Xi'an, CN)
- Hongling Chen (Xi'an, CN)
- Zhaoqi GAO (Xi'an, CN)
- Chuang Li (Xi'an, CN)
- Lijun Mi (Xi'an, CN)
- Jinmiao Zhang (Xi'an, CN)
- Qingzhen Wang (Xi'an, CN)
Cpc classification
International classification
Abstract
A model-driven deep learning-based seismic super-resolution inversion method includes the following steps: 1) mapping each iteration of a model-driven alternating direction method of multipliers (ADMM) into each layer of a deep network, and learning proximal operators by using a data-driven method to complete the construction of a deep network ADMM-SRINet; 2) obtaining label data used to train the deep network ADMM-SRINet; 3) training the deep network ADMM-SRINet by using the obtained label data; and 4) inverting test data by using the deep network ADMM-SRINet trained at step 3). The method combines the advantages of a model-driven optimization method and a data-driven deep learning method, and therefore the network has the interpretability; and meanwhile, due to the addition of physical knowledge, the iterative deep learning method lowers requirements for a training set, and therefore an inversion result is more reliable.
Claims
1. A model-driven deep learning-based seismic super-resolution inversion (SRI) method, comprising the following steps: 1) obtaining seismic data from an oil field and model data and using the seismic data and the model data to produce label data; 2) mapping each iteration of a model-driven alternating direction method of multipliers (ADMM) to each layer of a deep network, and learning proximal operators by using a data-driven method to complete a construction of a deep network ADMM-SRI; 3) training the deep network ADMM-SRI by using the label data; 4) inverting test data by using the deep network ADMM-SRI trained at step 3); 5) inverting the seismic data from the oil field using the deep network ADMM-SRI; and 6) selecting an oil or gas reservoir using the inverted seismic data from the oil field and performing exploration and/or development of the selected oil or gas reservoir in the oil field.
2. The model-driven deep learning-based seismic super-resolution inversion (SRI) method according to claim 1, wherein at step 2), a proximal form of the model-driven ADMM is shown as formula (1):
3. The model-driven deep learning-based seismic super-resolution inversion (SRI) method according to claim 2, wherein at step 2), the deep network ADMM-SRI comprises three stages, wherein in a k-th stage, the deep network is composed of three modules comprising the module r.sup.k, the module x.sup.k, and the module β.sup.k, wherein the module r.sup.k and the module x.sup.k are configured to calculate values of r.sup.k and x.sup.k by using learning operators; the module β.sup.k is configured to calculate a value of β.sup.k, and nodes of the three modules are connected via straight lines with a directionality.
4. The model-driven deep learning-based seismic super-resolution inversion (SRI) method according to claim 2, wherein a first learning module in formula (1) is as follow:
r.sup.k=Γ.sub.Θ.sub.
5. The model-driven deep learning-based seismic super-resolution inversion (SRI) method according to claim 2, wherein a second learning module in formula (1) is configured to learn a mapping relationship γ.sub.Θ.sub.
x.sup.k=γ.sub.Θ.sub.
6. The model-driven deep learning-based seismic super-resolution inversion (SRI) method according to claim 2, wherein for the third module in formula (1), only the parameter η needs to be determined in each iteration, η is considered as a weight in the deep network, and η is learned together with other parameters in the deep network from training data.
7. The model-driven deep learning-based seismic super-resolution inversion (SRI) method according to claim 1, wherein at step 2), the label data comprises the model data and field data, wherein: for the model data, a known velocity model is configured to perform a forward modeling to obtain seismic super-resolution data used as the label data; and for the field data, an acquisition procedure is as follows: the seismic data is first subjected to a denoising preprocessing and then subjected to a non-stationary correction to obtain stationary seismic data; a reflection coefficient profile is obtained by using an alternating iterative inversion method, and the reflection coefficient profile is filtered through a wide-band Gaussian or a Yu wavelet to obtain band-limited super-resolution data used as the label data.
8. The model-driven deep learning-based seismic super-resolution inversion (SRI) method according to claim 7, wherein step 3) of training the deep network ADMM-SRI by using the label data specifically comprises: preliminarily training the deep network by using the model data to enable parameters in the deep network to have an initial value; specifically obtaining the label data from the field data in different target areas, and performing a fine adjustment on the deep network by a transfer learning strategy, and configuring the deep network to invert the field data.
9. The model-driven deep learning-based seismic super-resolution inversion (SRI) method according to claim 8, wherein during the training of the deep network ADMM-SRI, a loss function is as follow:
E(Θ)=1/NΣ.sub.i=1.sup.N∥{circumflex over (r)}.sub.i−r.sub.i.sup.gt∥.sub.2.sup.2, (4) wherein a network parameter is Θ={Θ.sub.k.sup.g,Θ.sub.k.sup.f,η.sup.k}.sub.k=1.sup.N.sup.
10. The model-driven deep learning-based seismic super-resolution inversion (SRI) method according to claim 9, wherein in formula (4), an optimization is performed by a batch stochastic gradient descent (SGD) method, and a fixed number of iterations is set as a stop condition.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
DETAILED DESCRIPTION OF THE EMBODIMENTS
(14) In order to enable those skilled in the art to better understand solutions of the present disclosure, the technical solutions in embodiments of the present disclosure will be clearly and completely described below in conjunction with the drawings in the embodiments of the present disclosure. Apparently, the described embodiments are only some embodiments but not all embodiments of the present disclosure. All other embodiments obtained by those of ordinary skill in the art based on the embodiments of the present disclosure without creative work shall fall within the scope of protection of the present disclosure.
(15) It should be noted that the terms “first”, “second”, and the like in the description, claims and drawings of the present disclosure are used to distinguish similar objects, and are not necessarily used to describe a specific sequence or order. It should be understood that data used in this way may be interchanged under appropriate conditions so that the embodiments of the present disclosure described here can be implemented in a sequence other than those illustrated or described here. In addition, the terms “include” and “have” and any variation thereof are intended to cover non-exclusive inclusions. For example, a process, method, system, product, or device that includes a series of steps or units is not necessarily limited to those steps and units clearly listed, but may include other steps or units that are not clearly listed or are inherent to these processes, methods, products, or devices.
(16) The present disclosure will be further described below in detail with reference to the drawings.
(17) The present disclosure proposes a model-driven deep learning method to implement a seismic super-resolution inversion, which is called ADMM-SRINet. The method combines a model-driven alternating direction method of multipliers (ADMM) with a data-driven deep learning method to construct a deep network structure. Specifically, according to the method, each iteration of ADMM is mapped to each layer of a network, and proximal operators are learned by using the data-driven method. All parameters such as a regularization parameter and a transformation matrix in the network may be implicitly learned from a training data set, and are not limited to a form of regularization term. In addition, for complex field data, the present disclosure designs a set of processes for obtaining label data and a novel solution for network training, so ADMM-SRINet may be used to better invert the field data. Finally, the network is configured to invert synthetic and field data, which verifies the effectiveness of the present disclosure.
(18) 1. ADMM-Based Seismic Super-Resolution Inversion
(19) Based on a conventional convolution model, seismic records may be modeled by using the following mathematical framework:
y=Wr, (1)
(20) y∈R.sup.n represents observed data, W∈R.sup.n×m is a convolution matrix composed of seismic wavelets, and r∈R.sup.m is a super-resolution result to be solved. A main objective of the seismic super-resolution inversion is to optimize an objective function in the following formula:
(21)
(22) D represents a sparse transformation matrix, λ is a regularization parameter, and ∥•∥.sub.P represents a norm of l.sub.p(0≤p≤1). In order to solve formula (2), various iterative optimization algorithms such as an alternating direction method of multipliers (ADMM) and an iterative shrinkage threshold algorithm (ISTA) have been proposed. In the present disclosure, ADMM is adopted.
(23) ADMM is also known as a Douglas-Rachford splitting algorithm that may be used to split an objective function into multiple sub-problems and then perform alternating solution. ADMM may be interpreted by using a thought of augmented Lagrangian. First an auxiliary variable x is introduced, and formula (2) may be written as the following form:
(24)
(25) An augmented Lagrangian form of formula (3) is as follow:
(26)
(27) where ρ is a penalty parameter, and α is a dual variable. Based on three variables in formula (4), formula (4) is split into the following three sub-problems:
(28)
(29) where
(30)
and η is an updated parameter. A first formula and a third formula in formula (5) are easy to solve, but a solution of a second formula is challenging, specially under a condition that the regularization term is nonconvex, it is difficult for researchers who are not engaged in algorithm optimization to solve the non-convex problem. Generally, when ∥Dx∥.sub.p=∥Dx∥.sub.1 and D is an orthogonal matrix, a solution of each sub-problem in formula (5) is as follow:
(31)
(32) where S.sub.λ/ρ is a soft threshold function with a threshold of Alp. Other threshold functions may be selected to replace the soft threshold function.
(33) ADMM described above is a model-driven method, where the regularization term λ, the parameter ρ, the sparse matrix D, and some other hyper-parameters in ADMM need to be determined in advance. In addition, ADMM needs to be subjected to multiple iterations to achieve a satisfied result, which brings a big challenge to an inversion of high-dimensional data. Moreover, the non-orthogonal matrix D and 0≤p<1 make formula (5) difficult to be solved.
(34) 2. Model-Driven ADMM-SRINet
(35) In order to solve the limitation of ADMM, a model-driven deep network is designed to implement a seismic super-resolution inversion, which is called ADMM-SRINet. In order to introduce the proposed network structure, formula (5) is written as the following proximal form:
(36)
(37) where η is a parameter varying with the number of iterations, and prox.sub.ρf(•) and prox.sub.ρg (•) are proximal operators and are defined as follow:
(38)
(39) and
(40)
(41) For an input variable {circumflex over (z)},
(42)
and g({circumflex over (z)})=λ∥D{circumflex over (z)}∥.sub.p. By observing formula (7), it can be seen that the proximal operators prox.sub.ρf (•) and prox.sub.ρg(•) are the keys to solve the inverse problems and may be replaced with other operators such as a denoising operator. In the present disclosure, based on the inspiration of the strong learning capacity of the deep learning, a learning operator is used to replace the proximal operator according to Adler's work in 2018, wherein parameters in the learning operator may be obtained through training. Therefore, a relatively optimal solution may be obtained for the proximal problem under a relatively small number of iterations, which avoids the determination of some parameters. Although there is no explicit expression for learning the proximal operator, a universal approximation property of a neural network ensures that these operators may be arbitrarily approximated. The following is a detailed description of ADMM-SRINet.
(43) A. First Learning Operator
(44) Although there is an analytical solution for the first proximal problem in formula (7), the calculation of the matrix increases the calculation cost, and the selection of the parameters increases the difficulty of solution. Therefore, different form Yang's work in 2018, one residual convolutional network block is used to replace the first proximal operator prox.sub.ρf(•) so as to learn a mapping relationship between r.sup.k and (x.sup.k-1−β.sup.k-1), which is expressed by the following formula:
r.sub.k=Γ.sub.Θ.sub.
(45) where Γ.sub.Θ.sub.
(46) B. Second Learning Operator
(47) The second proximal problem in formula (7) is usually non-convex, and it is difficult to select an appropriate normal form to obtain an optimal result, and therefore it is a novel way to solve the limitation of the conventional method by using the currently rapidly developed deep learning method. In order to design a network structure to learn the proximal operator prox.sub.ρg(•), the second formula in formula (5) is rewritten as the following form:
(48)
(49) where p.sup.k=r.sup.k+β.sup.k-1, and F represents a nonlinear sparse function. Based on theorem 1 in the work of Zhang, et al. in 2018, the following approximation expression may be obtained:
∥F(x)−F(p.sup.k)∥.sub.2.sup.2≈υ∥x−p.sup.k∥.sub.2.sup.2, (12)
(50) where p.sup.k and F(p.sup.k) are assumed to be mean values of x and F(x). Therefore, formula (11) may be rewritten as follow:
(51)
(52) where g′(•)=ç∥F(•)∥.sub.p, and ç=υλ. By using the residual convolutional network block, a solution of formula (13) may be obtained through one learning operator, which is expressed as follow:
F(x.sup.k)=Λ.sub.Θ.sub.
(53) where ΛΘ.sub.k.sub.
x.sup.k=F.sup.H.sup.
(54) where similarly the learning operator is used to transform function F and its inverse function, and F.sup.k and F.sup.H.sup.
x.sup.k=p.sup.k+Q.sup.k(F.sup.H.sup.
(55) where a learning operator R.sup.k=Q.sup.koJ.sup.k is used to extract a lost high frequency component from x.sup.k, that is, w.sup.k=R.sup.k(x.sup.k), so under a noiseless condition, x.sup.k=p.sup.k+w.sup.k.
(56) By observing formula (16), it can be seen that if operators Q.sup.kF.sup.H.sup.
(57) Based on the above derivation,
x.sub.k=γ.sub.Θ.sub.
(58) where Θ.sub.k.sup.g represents a network parameter in the kth iteration of the network.
(59) C. Network Structure of ADMM-SIRNet
(60) Based on the above descriptions of A and B, it can be seen that the proximal operator in formula (7) may be replaced with a residual convolutional network. For the third part in formula (7), only one parameter η needs to determine in each iteration. η may be considered as a weight in the network, which is learned together with other parameters in the network from training data. Finally, a complete network structure of ADMM-SRINet is shown in
(61) D. Network Optimization of ADMM-SRINet
(62) In order to obtain all parameters in ADMM-SRINet, the following function is minimized:
(63)
(64) where a network parameter is Θ={Θ.sub.k.sup.g,Θ.sub.k.sup.f,η.sup.k}.sub.k=1.sup.N.sup.
(65) A material basis of the present disclosure is a seismic data volume, and a trace-by-trace processing method is used. Specific steps may refer to a process framework in
(66) step 1: each iteration of a model-driven alternating direction method of multipliers (ADMM) is mapped to each layer of a deep network, and proximal operators are learned by using a data-driven method to complete construction of a deep network ADMM-SRINet (see
(67) step 2: label data used to train the deep network ADMM-SRINet is obtained;
(68) step 3: the deep network ADMM-SRINet is trained by using the obtained label data; and
(69) step 4: test data is inverted by using the deep network ADMM-SRINet trained at step 3).
(70) Effectiveness Analysis:
(71) 1. Synthetic Data Example
(72) Due to a relatively low signal to noise ratio of field seismic data, it is usually difficult to recover an underground full-band reflection coefficient. In contrast, an inversion result of a band-limited reflection coefficient is more reliable. Therefore, in the present disclosure, the super-resolution inversion method is mainly used to invert a band-limited reflection coefficient from seismic data. First, in order to train the proposed network ADMM-SRINet, a Marmousi II velocity model was used to generate a reflection coefficient profile, and then the reflection coefficient module was convoluted with a Yu wavelet (a dotted line in
(73) In order to verify the effectiveness of ADMM-SRINet, the network was used to invert a model test data set. The test set was generated by convoluting the data outside the dotted block in
(74) 2. Field Seismic Data Example
(75) One section of a post-stack three dimensional data volume of a certain oil field was selected to test the method of the present disclosure. Due to the complex field seismic data, in order to obtain a relatively good inversion result, an inversion process was designed, as shown in
(76) In order to enhance the reliability of the label data, super-resolution data was obtained by using an alternating iterative reflection coefficient inversion method proposed by Wang in 2016, so not only relatively reliable label data may be obtained, but also wavelet information may be obtained. For an inversion of the field data, similarly a band-limited reflection coefficient was inverted. In order to enable the obtained inversion result to have a higher resolution, a wide-band Gaussian wavelet shown in
(77) In conclusion, according to the present disclosure, each iteration of ADMM is mapped into each layer of a network, and the proximal operators are learned by using a data-driven method. The method fully combines advantages of a model-driven alternating direction method of multipliers (ADMM) and a data-driven deep learning method, avoids the design of regularization terms and then implements a fast calculation of high-dimensional data. In addition, the present disclosure designs a set of process for obtaining label data and a novel solution for network training, and mainly uses model label data and field label data to train the deep network through a transfer learning strategy, and thus ADMM-SRINet can be used to better invert field data. Finally, the network is used to invert synthetic and filed data, which verifies the effectiveness of the present disclosure.
(78) The above content is only to illustrate the technical ideas of the present disclosure, and cannot be used to limit the scope of protection of the present disclosure. Any changes made on the basis of the technical solutions based on the technical ideas proposed by the present disclosure shall fall within the scope of protection of the claims of the present disclosure.