Guided bayesian experimental design
09720130 · 2017-08-01
Assignee
Inventors
- Hugues A. Djikpesse (Cambridge, MA)
- Mohamed-Rabigh Khodja (Quincy, MA)
- Michael David Prange (Somerville, MA)
US classification
- 1/1
Cpc classification
G01V11/00
PHYSICS
G01V99/00
PHYSICS
International classification
G01V11/00
PHYSICS
G06F17/18
PHYSICS
Abstract
A Bayesian methodology is described for designing experiments or surveys that are improved by utilizing available prior information to guide the design toward maximally reducing posterior uncertainties in the interpretation of the future experiment. Synthetic geophysical tomography examples are used to illustrate benefits of this approach.
Claims
1. A method for characterizing a subterranean formation, the method comprising: selecting model parameters of a linear model that predicts physical observations of the formation; determining a prior model probability density relating to the model parameters of the linear model, the prior model probability density including (i) a mean prior model, and (ii) a first covariance matrix describing uncertainty around the mean prior model; determining a second covariance matrix relating to the model parameters of the linear model and describing uncertainty around anticipated observation noise; and iteratively adding a select candidate physical observation to a set of physical observations of the formation as predicted by the linear model based on D-optimality augmented criteria using a processor, which involves determining a sensitivity matrix for the model parameters and the set of physical observations based on a determinant derived from the first covariance matrix and components of the second covariance matrix that are associated with the select candidate physical observation; and configuring survey equipment that characterizes the formation using the set of physical observations.
2. The method of claim 1, further comprising assigning the physical model parameters to a probability density function.
3. The method of claim 1, wherein each physical observation is associated with multiple measurements.
4. The method of claim 3, further comprising processing k measurements at a time for a physical observation with k associated measurements.
5. The method of claim 1, wherein the determinant is further derived from components of the second covariance matrix that are associated with a number of base physical observations.
6. The method of claim 1, wherein at least one select candidate physical observation is iteratively added to the set of physical observations of the formation until a stop condition is satisfied.
7. The method of claim 1, wherein the survey equipment comprises an array of seismic receivers.
8. Apparatus for characterizing a subterranean formation, the apparatus comprising: a memory storing comprising physical model parameters of a linear model that predicts physical observations of the formation; and a processor which is configured to: determine a prior model probability density relating to the model parameters of the linear model, the prior model probability density including (i) a mean prior model, and (ii) a first covariance matrix describing uncertainty around the mean prior model, determine a second covariance matrix relating to the model parameters of the linear model and describing uncertainty around anticipated observation noise, and iteratively add a select candidate physical observation to a set of physical observations of the formation as predicted by the linear model based on D-optimality augmented criteria, which involves determining a sensitivity matrix for the model parameters and the set of physical observations based on a determinant derived from the first covariance matrix and components of the second covariance matrix that are associated with the select candidate physical observation; wherein the set of physical observations are used to configure survey equipment that characterizes the formation.
9. The apparatus of claim 8, wherein the determinant is further derived from components of the second covariance matrix that are associated with a number of base physical observations.
10. The apparatus of claim 8, wherein at least one select candidate physical observation is iteratively added to the set of physical observations of the formation until a stop condition is satisfied.
11. The apparatus of claim 10, wherein the stop condition involves at least one of: i) a condition related to the determinant, ii) a condition related to the number of candidate physical observations added to the set of physical observations and to the total number of model parameters of the linear model, and iii) a condition related to the number of candidate physical observations added to the set of physical observations and a predefined maximum number.
12. The apparatus of claim 8, wherein the survey equipment comprises an array of seismic receivers.
13. The method of claim 6, wherein the stop condition involves at least one of: i) a condition related to the determinant, ii) a condition related to the number of candidate physical observations added to the set of physical observations and to the total number of model parameters of the linear model, and iii) a condition related to the number of candidate physical observations added to the set of physical observations and a predefined maximum number.
Description
BRIEF DESCRIPTION OF THE FIGURES
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
DETAILED DESCRIPTION
(12) A general arrangement of a vertical seismic profile (VSP) survey is shown in
(13)
(14)
(15)
In step 306, n is incremented. Then, for n=1,
(16)
as indicated by step 308. Steps 304, 306 and 308 are repeated until convergence as determined in step 310. For example, in a subsequent iteration for n=1, find g.sub.2 that maximizes
(17)
and then for n=2,
(18)
and find g.sub.3 that maximizes
(19)
Examples of stop 312 conditions include: adding a new observation to the experiment no longer results in significant improvement in |C.sub.n+1.sup.−1|/|C.sub.n.sup.−1|; when the number of added measurements exceeds the total number of model parameters; and when the number of added measurements exceeds a predefined maximum number. After completion, the optimal design is the set of observations that would maximally impact the resolution of the model parameters given the information available on the model parameters and the anticipated observation noise prior to collecting the measurement.
(20) In view of the description above, it will be appreciated that the algorithm is advantageously generic in the sense that it can be applied with any physical measurement or observation, i.e., it is not limited to seismic and acoustic surveys. However, at least with regard to seismic and acoustic surveys the algorithm provides the advantage of utilizing prior model information which often exists but might not otherwise be used. For instance, in geophysical tomography, the prior mean model and associated covariance matrix could come from surface seismic data interpretation when one is considering 3D vertical seismic profile acquisition to refine a particular area of the subsurface model. When one is performing a real-time survey design, the prior information on the model could come from the interpretation of the already acquired measurements.
(21) It should be noted that although the algorithm has been described for the case of single measurement observations and uncorrelated noise, it is also applicable when considering observations with multiple measurements and correlated noise by changing the function to maximize for finding g.sub.n and by changing the formula used to update the base experiment matrix C.sub.n. In the case where each observation may be associated with multiple measurements, the observation selection algorithm considers k measurements at a time for an observation with k associated measurements. To compare the numerical advantage of performing one-time rank-k updates versus over performing k consecutive rank-one updates, let
Γ=[Γ.sub.1Γ.sub.2 . . . Γ.sub.k].sup.T (26)
be the matrix whose rows Γ.sub.i, 1≦i≦k, are the sensitivity kernels of the relevant data stations. For a diagonal data covariance matrix, i.e., for
(22)
It follows that for a one-time rank-k augmentation
C.sub.n+k.sup.−1=C.sub.n.sup.−1+Γ.sup.TSΓ, (29)
from which it is straightforward show that
(23)
This expression reduces for k=2 to
(24)
Examining the operations count for a one-time rank-k augmentation one has to sum the operations occurring in four computation steps: 1. The product of a typically large M×M matrix with a M×k matrix requires M×M×k multiplications and M×k×(M−1) additions. This step requires O(M.sup.2k) operations. 2. The product of a k×M matrix by a M×k matrix requires k×k×M multiplications and k×k×(M−1) additions. This requires O(k.sup.2M) operations. 3. The determinant of a typically small k×k matrix requires O(k.sup.3) operations. 4. The product of k+1 scalars requires O(k) operations.
(25) Therefore the total cost of this one-time rank-k augmentation procedure is
O(M.sup.2k)+O(k.sup.2M)+O(k.sup.3)+O(k)≈O(M.sup.2k) for k<<M. (33)
Now, for k consecutive rank-one updates
(26)
An analytic expression |C.sub.n+k.sup.−1|/|C.sub.n.sup.−1| for a given k, is
(27)
The explicit form of Eq. 35 for an arbitrary k is somewhat cumbersome, but for k=2 it can be seen that it is identical to Eq. 32. Examining the operations count for k consecutive rank-one updates, one has to perform k(M.sup.2+M+2)+k multiplications and k(M(M−1)+(M−1)+1) additions, yielding the operations count estimate
O(M.sup.2k) for k<<M. (37)
Comparing the operations counts for both update approaches, there is no advantage to using one over the other. However, in comparing the implementation complexity of Eq. 35 versus Eq. 30, one might prefer the simplicity of the former over the latter.
(28) In the most general case the data covariance matrix, C.sub.D, is a symmetric, positive definite matrix; it is conveniently written in block form as
(29)
wherein (C.sub.D).sub.n is the covariance matrix of the base experiment data, σ.sub.n+1.sup.2 is the variance of the data measurement that corresponds to the new candidate observation, and c.sub.n+1 is the vector whose components are the covariance terms of this measurement. Using the formula for the inverse of a block matrix (G. H. Golub and C. F. Van Loan. Matrix Computations. Johns Hopkins, Baltimore, Md., USA, 1996),
(30)
Substituting (39) into (8) yields
(31)
Upon using the Sherman-Morrison formula described by Golub and Loan, Eq. (40) reduces to
(32)
The repeated use of the matrix determinant lemma (D. A. Harville. Matrix Algebra from a Statistician's Perspective. Springer-Verlag, New York, N.Y., USA, 1997) and the Sherman-Morrison formula yields for Eq. 41
(33)
Note that the Woodbury formula described by Harville, a rank-k-generalization of the Sherman-Morrison result, could be used to calculate A.sup.−1. This yields
A.sup.−1=(C.sub.n.sup.−1=G.sub.n.sup.TBG.sub.n).sup.−1=C.sub.n−C.sub.nG.sub.n.sup.T(B.sup.−1+G.sub.nC.sub.nG.sub.n.sup.T).sup.−1G.sub.nC.sub.n. (37)
However, from a computational point of view this is not very helpful as one would still have to calculate the inverse of two large matrices. To calculate |A| the generalized matrix determinant lemma can be used, which yields
|A|=|C.sub.n.sup.−1+G.sub.n.sup.TBG.sub.n|=|C.sub.n.sup.−1∥I+BG.sub.nC.sub.nG.sub.n.sup.T|. (48)
wherein I is the identity matrix. With these results, Eqs. (46) and (48), the objective function can be expressed as
(34)
wherein A is given by Eq. 42, B is given by Eq. 43, h is given by Eq. 44, and k is given by Eq. 45. It can be shown that when the data covariance matrix is diagonal, i.e., when c.sub.n+1=0, Eq. 49 reduces to Eq. 16. This result, Eq. 49, would clearly be computationally expensive to implement, but may be an acceptable cost when the data measurements are not independent from one another.
(35)
(36)
(37) While the invention is described through the above exemplary embodiments, it will be understood by those of ordinary skill in the art that modification to and variation of the illustrated embodiments may be made without departing from the inventive concepts herein disclosed. Moreover, while the preferred embodiments are described in connection with various illustrative structures, one skilled in the art will recognize that the system may be embodied using a variety of specific structures. Accordingly, the invention should not be viewed as limited except by the scope and spirit of the appended claims.