METHOD FOR RECONSTRUCTING AT LEAST ONE TRACE IN A SEISMIC IMAGE
20230066911 · 2023-03-02
Inventors
- Prashanth NADUKANDI (Madrid, ES)
- Pablo Enrique VARGAS MENDOZA (Madrid, ES)
- Santiago FERNÁNDEZ PRIETO (Madrid, ES)
- German OCAMPO BOTERO (Madrid, ES)
- Wenyi HU (Sugar Land, TX, US)
- Shirui WANG (Houston, TX, US)
- Pengyu YUAN (Houston, TX, US)
Cpc classification
G01V1/345
PHYSICS
G01V2210/1429
PHYSICS
International classification
Abstract
The present invention is related to a method for reconstructing at least one trace in a seismic image of a common receiver and time domain, the image comprising traces in time domain with seismic data and one or more traces to be reconstructed. A first aspect of the invention is a method that is characterized by a specific use of a convolutional neural network trained under an unsupervised learning approach with a modified receptive field. A second aspect of the invention is a deblending method based on the use of a reconstructing method according to the first aspect of the invention applied to a denoising step of a deblending process allowing a very effective data acquisition while keeping a high quality output data sets after being processed according to the first and/or second aspects of the invention.
Claims
1. A method for reconstructing at least one trace in a seismic image, the method comprising the steps of: a) providing a seismic image having seismic traces extracted from a data set acquired via a seismic survey, wherein the seismic image includes at least one seismic trace with data that is unavailable from the acquired data set; b) training a convolutional neural network to predict values for the unavailable data of the at least one trace, the convolutional neural network includes at least one layer having a kernel function configured to evaluate a bounded domain that defines a blind-trace receptive field, wherein the blind-trace receptive field encompasses data corresponding to a plurality of traces in the seismic image that are adjacent to, but exclude, the at least one trace having unavailable data; c) predicting values for the unavailable data of the at least one trace by inputting the seismic image into the trained convolutional neural network; d) reconstructing the at least one trace having unavailable data using the data values predicted in step c); and e) generating a reconstructed seismic image with the at least one trace reconstructed in step d).
2. The method according to claim 1, wherein the convolutional neural network is a U-net.
3. The method according to claim 1, wherein the at least one layer is a down-sampling layer.
4. The method according to claim 1, wherein the blind-trace receptive field of the kernel function is determined by the following processing steps: limiting the receptive field to one adjacent side of the at least one trace having unavailable data, the one adjacent side being determined based on a direction of the at least one trace having unavailable data; generating a first copy of the seismic image and a second copy of the seismic image, the second copy of the seismic image being rotated 180° with respect to the first copy of the seismic image; inputting the first copy and the second copy of the seismic image into the convolutional neural network with the one-adjacent-sided receptive field resulting in two output images that are subsequently combined to form a single image.
5. The method according to claim 4, wherein the at least one layer with a blind-trace receptive field is a down-sampling layer and wherein any layer of the convolutional neural network includes an output for outputting a feature map and, wherein: before inputting a feature map, output of a previous layer or the image if the current layer is the first layer, into the current layer, the feature map is padded by adding rows of zeros at the end of the feature map located at one side of the trace; carrying out a convolution operation; cropping out the same number of rows previously added wherein cropping the output feature map is carried out on the side opposite to the side on which rows were previously added.
6. The method according to claim 2, wherein training the convolutional neural network includes using a converging criterion based on an approximation error estimation E.sub.s for measuring an interpolation loss when predicting the reconstructed trace, and the approximation error estimation E.sub.s being determined as a linear combination of a misfit loss and a regularization loss, the regularization loss being determined by the following steps: determining a main energy area of the seismic image; calculating a norm of the reconstructed seismic image in the frequency domain limited to an area outside of the main energy area based on a predetermined norm.
7. The method according to claim 6, wherein the linear combination determining the approximation error estimation is:
E.sub.S=∥misfit loss∥+αƒregularization loss∥ being α a positive weighting value and ∥⋅∥ the predetermined norm.
8. The method according to claim 7, wherein the misfit loss is determined as a difference between the seismic image and the reconstructed seismic image after removing the at least one reconstructed trace.
9. A method for deblending seismic data in a receiver domain, the method comprising: deploying a plurality of n.sub.s acoustic sources in the upper surface of the reservoir domain wherein the n.sub.s acoustic sources are grouped in B groups σ.sub.i, l=1 . . . B, of acoustic sources, each acoustic source being in only one group σ.sub.i of sources and at a location x.sub.i.sup.s, i=1 . . . n.sub.s and, deploying a plurality of n.sub.r, acoustic receivers in the upper surface of the reservoir domain at a location x.sub.j.sup.r, j=1 . . . n.sub.r; a) for each group σ.sub.i, l=1 . . . B of acoustic sources, each acoustic source is shot with a random delay time τ.sub.li and the response in the acoustic receivers stored in a data structure that may be represented by P.sub.ljk.sup.b=(x.sub.i.sup.s, x.sub.i.sup.r, t.sub.k−τ.sub.li; wherein t.sub.k is the k.sup.th time sample in the time domain; b) calculating, for each group σ.sub.l, l=1 . . . B, the Fourier-transform Π.sup.b(:,:,ω.sub.k)=F{P.sub.ljk.sup.b} wherein ω.sub.k is the k.sup.th frequency and “:” denoting variables depending in index i or index j; c) for each frequency ω.sub.k determining Π.sup.LS(:,:,ω.sub.k)=Γ*Π.sup.b(:,:,ω.sub.k) wherein Γ* is
Γ*=Γ.sub.k.sup.HD being D a diagonal matrix and Γ.sub.k.sup.H the conjugate transpose of Γ.sub.k the blending matrix that may be calculated from the random delay times τ.sub.li as
10. The method for deblending seismic data according to claim 9, wherein D=(Γ.sub.kΓ.sub.k.sup.H).sup.−1.
11. The method for deblending seismic data according to claim 9, wherein D=I, wherein I denotes the identity matrix.
12. The method for deblending seismic data according to claim 9, wherein each source is shot once and wherein for all m≠n, σ.sub.m∩σ.sub.n=Ø.
13. A non-transitory computer program product stored on a computer-readable medium and comprising computer-implementable instructions, which, when executed by a computer, cause the computer to carry out the method according to claim 1.
14. A computer system having a processor and a non-transitory computer-readable medium storing computer-executable instructions which, when executed by the processor, cause the processor to carry out the method according to claim 1.
Description
DESCRIPTION OF THE DRAWINGS
[0091] These and other features and advantages of the invention will be seen more clearly from the following detailed description of a preferred embodiment provided only by way of illustrative and non-limiting example in reference to the attached drawings.
[0092]
[0093]
[0094]
[0095]
[0096]
[0097]
[0098]
[0099]
[0100]
[0101]
[0102]
[0103]
DETAILED DESCRIPTION OF THE INVENTION
[0104] As will be appreciated by one skilled in the art, aspects of the present invention may be embodied method that may be implemented as a computer program product, at least those parts manipulating the acquired seismic data.
[0105] In order to show an specific embodiment of the disclosed methods in a detailed manner, let represent the 1-D Fourier transformation along the time axis, that is, for what has been named a trace,
.sup.−1 its inverse and f.sub.θ(⋅) is a denoising convolutional neural network (CNN) with parameters θ. Further, let C denote the array concatenation operator which takes arrays and stacks them into a single higher-dimensional array, e.g. Π=CΠ.sub.k. With no unblended data available, a model that minimizes the unsupervised blending loss function is trained, being the loss function
[0106] and ={x.sup.(1), . . . , x.sup.(N)}, where the input x=.sup.−1CΓ.sub.m.sup.HΠ.sub.k.sup.b, is the pseudo-deblended data in the common receiver domain. f.sub.θ(x) gives the prediction of the deblended data. The prediction of the model is blended again by the known blending matrix Γ. This result is the blended prediction: Γ.sub.m(
f.sub.θ(x.sup.(k))).sub.m. The term
.sup.−1CΓ.sub.m.sup.HΓ.sub.m(
f.sub.θ(x.sup.(k))).sub.m is the pseudodeblended output of the model's prediction. The mean absolute error is taken from the pseudodeblended output of the model's prediction and the original pseudo-deblended data, i.e. the input x. The action of the blending followed by the pseudodeblending operator
.sup.−1CΓ.sub.m.sup.HΓ.sub.m(
(⋅)).sub.m can be implemented efficiently and directly in the time domain by first combining the traces according to the blended-acquisition time dithering code, then each bended shot-gather is copied a number of times equal to the corresponding source group size, and finally each of these copies is time-shifted (dithering decoded) to undo the delays introduced in the field. By adopting this blending loss, the deblending task is accomplished in a unsupervised way with the blended data and the blending matrix.
[0107] Direct minimization on this blending loss with a traditional deep CNN leads to identity mapping. It will produce the pseudo-deblended result which is exactly the least square solution. This phenomenon could be observed in
[0108] To avoid this local minimum, a strong constraint for the model by making it trace-wise blind is proposed. According to this proposal, the coherent signal of the traces are essentially reconstructed from their adjacent traces, while not looking at themselves in the input. Thus for a specific trace, there would be no chance for its blending noise to be mapped from input to output. For those noises in its adjacent area, the model automatically neglect them due to their discontinuity and the convergence of the blending noise.
[0109] The illustration is shown in
[0110] According to this embodiment, the blind-trace convolutional neural network is constructed for the deblending task.
[0111] The network structure is shown in
[0112] In the blind-trace U-net, the receptive field of each layer is strictly restricted to the upper half area for each row so that the model can extract the coherent features for each trace based on its left or right adjacent area in the original input respectively.
[0113] After the blind-trace U-net, the two output feature maps are cropped at the bottom and padded at the top such that the target traces can be excluded from its receptive field to fulfill the “blind-trace” purpose.
[0114] Then, they are rotated back (270° and 90° accordingly) and concatenated, followed by two consecutive 1×1 convolutional layers that integrate the feature maps and squeeze the channel size to 1.
[0115] In
[0116] According to a preferred embodiment that may be applied to any of the disclosed examples, before the padding and cropping operations, the trace is buried under the patches, and at the end of the process, the trace will be excluded from the patches. For the edges, there is only one side of the patch that is informative and zeros are padded on the other side.
[0117] The image is processed only once and the network gives the predictions for all traces at the same time.
[0118] Specifically, to constrain the receptive field within the upper half for all rows in the blind-trace U-net, the convolutional layers and down-sampling layers are changed as shown in
[0119] Convolution: The feature maps are padded with zeros before each convolutional layer. Given the size of the filter k×k, └k/2┘ lines are padded on the top of the feature maps, and crop the └k/2┘ lines at the bottom after the convolution.
[0120] Max-pooling: The feature maps are padded on the top with one line and crop one at the bottom before max-pooling. There is no need to change the up-sampling layers since it does not affect the receptive filed after the modification for the down-sampling layers.
[0121] With the aforementioned changes, for a single trace in the feature maps, it takes the information purely from itself and the area above in the input. No information below it can leak into the result.
[0122] According to this embodiment, the rotated inputs are projected to 32 feature maps in the first layer. There are 4 contracting blocks and 4 expanding blocks in the blind-trace U-net, each of which consists of two consecutive modified convolutional layers and a max-pooling/up-sampling layer. The feature maps are doubled in the last convolutional layer of each contracting block and halved correspondingly in the expanding blocks.
[0123] Each blind-trace convolutional layer in the encoder is replaced by a blind-trace residual block (i.e. a residual block with all convolutional layers modified). Batch Normalization is adopted before each activation in the decoder.
[0124] The last two 1×1 convolutional layers project the concatenated 64 feature maps to 32 and 1 respectively. It has been used ReLU activation in the intermediate convolutional layers and leaky ReLU for the last 1×1 convolutional layer.
[0125] The proposed blind-trace network with blending loss combines the merits of both the conventional filter-based method and the inversion-based method.
[0126] The large number of the weights in the U-Net endows the network to achieve a complex non-linear filter to extract information from the coherent signal, meanwhile, for the low-SNR area where filters could not obtain any coherent information, the minimization of the blending loss reconstructs the coherent signal underneath through nonlinear inversion. This is a one-stage deblending framework and does not require much exhaustive and meticulous hyperparameter tuning.
[0127] Regarding the experiment, three different types of blending schemes are used to demonstrate the performance of the proposed method. The synthetic data examples can be seen in
[0128] “alternate” blending. In this blending strategy, two consecutive shots are blended using short and random dithering time. After pseudo-deblending, the noise will cover the whole common receiver gather as shown in
[0129] 200 unblended shots are collected with the size (1600(receiver numbers)×750(time samples)). Random delays in the range of (0, 1]s are added to even shots and blended with the odd shots. The sampling period is 6 ms.
[0130] “half” blending. The second half of the shot gathers are shifted with relatively long delays and added to the first half. Noises in the pseudo-deblended common receiver gathers concentrate in the two corners of the image as shown in
[0131] “continuous” blending. Simultaneous source of streamer acquisition is simulated in this experiment. A BP2004 velocity model is used for generating 200 consecutive unblended shots.
[0132] (BP2004 denotes the 2004 British Petroleum Velocity Benchmark. https://software.seg.org/datasets/2D/2004_BP_Vel_Benchmark/eage_abstract.pdf)
[0133] The shots have been recorded every 5 to 6 seconds continuously. There are 959 receivers for each record and 3334 samples along the time axis (dt=6 ms, hence T≈20 s). Approximately 4 shots are blended at different locations in the common receiver gathers. The noise level, in this case, is much higher than the other blending schemes as shown in
[0134]
[0135] Results of One-Stage Deblending
[0136] In order to avoid massive matrix multiplication in the f-x domain in loss computation, the blending and pseudo-deblending is performed in the time-space by applying dithering codes to the shots, followed by summation and cropping. In the first stage, the weights and bias are randomly initialized and trained with the Adam optimizer.
[0137]
[0138]
[0139] “Half” blending scheme introduces more challenges for deblending models. Due to the concentration of the noises and long dithering time, the weak amplitudes in late arrivals are severely contaminated by strong noises from the early arrivals. As can be seen in
[0140] In “continuous” blending scheme, more unblended shots are collected within one blended shot. The noisy area enlarges to the whole image, and the signal is covered by multiple layers of noises. Less coherent information can be attained by the model.
[0141] However, the one-stage deblending can still give a very good result as demonstrated in
TABLE-US-00001 Blending scheme Pseudo-deblended Training-stage Tuning-stage “alternate” 0.5178 0.1594 0.1141 “half” 0.0674 0.0169 0.0072 “continuous” 0.3250 0.1176 0.0789