Process parameter prediction using multivariant structural regression
20210182597 · 2021-06-17
Inventors
Cpc classification
G06F17/18
PHYSICS
International classification
G06F17/18
PHYSICS
Abstract
Multivariant feature extraction is used for training volumes or 2D images, (real or synthetic) coupled to process (effective) values probably obtained from direct simulation. These features are coupled with machine learning/regression algorithms to make a predictive model for the effective property. This model can then be used on a real geometry of a sample for effective parameter prediction.
Claims
1. A method for determining parameters of a sample, comprising: performing parameter estimation process by creating a prediction model from images; and making final predictions using real geometry of a sample and the prediction model.
2. The method of claim 1, wherein the prediction model is created from synthetic images.
3. The method of claim 1, wherein the prediction model is created from actual images.
4. The method of claim 1, wherein the prediction model is generated using multivariant regression.
5. The method of claim 1, wherein the prediction model is generated by computing process parameters from the images, segmenting the images into objects, extracting features from the images and computing statistics across the objects of the images.
6. The method of claim 5, wherein the prediction model is created by determining feature vectors from statistics determined from the images and then creating the prediction model by multivariate regression.
7. The method of claim 1, wherein the prediction model is created by feeding images into a convolutional neural network which maps from image area to a variable.
8. The method of claim 1, wherein the prediction model is generated using machine learning.
9. The method of claim 1, wherein the final predictions performed by extracting feature vectors from the sample and employing the prediction module to determine the parameters of the sample.
10. The method of claim 1, wherein the sample is from mining and/or oil/gas extraction.
11. The method of claim 1, wherein the sample is from an additive manufacturing process.
12. The method of claim 1, wherein the parameter estimation process is performed on 3D image datasets.
13. The method of claim 12, wherein the parameter estimation process is performed on 3D image datasets from an x-ray microscope.
14. The method of claim 1, wherein the final predictions are made from 2D image datasets.
15. The method of claim 14, wherein the final predictions are made from 2D electron or optical microscope image datasets.
16. A system for determining parameters of a sample, comprising: a machine learning application performing parameter estimation process by creating a prediction model from images; and the machine learning application making final predictions using real geometry of a sample and the prediction model.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0021] In the accompanying drawings, reference characters refer to the same parts throughout the different views. The drawings are not necessarily to scale; emphasis has instead been placed upon illustrating the principles of the invention. Of the drawings:
[0022]
[0023]
[0024]
[0025]
[0026]
[0027]
[0028]
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
[0029] The invention now will be described more fully hereinafter with reference to the accompanying drawings, in which illustrative embodiments of the invention are shown. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art.
[0030] As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items. Further, the singular forms and the articles “a”, “an” and “the” are intended to include the plural forms as well, unless expressly stated otherwise. It will be further understood that the terms: includes, comprises, including and/or comprising, when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. Further, it will be understood that when an element, including component or subsystem, is referred to and/or shown as being connected or coupled to another element, it can be directly connected or coupled to the other element or intervening elements may be present.
[0031] Important macroscopic or effective parameters of a sample include but are not limited to the sample's permeability, diffusivity, elastic moduli, electrical characteristics, and single or multiphase transport processes. These parameters are frequently calculated from 3D tomographic images using computational simulation of partial differential equations.
[0032]
[0033] A 3D volume 108 of a sample, such as one created by tomographic reconstruction from x-ray projections through the sample, is used to create a full physics simulation 110. This simulation 110 is then used to predict the sample's process parameters of interest such as the sample's permeability 112.
[0034] Performing a full physics simulation is extremely computationally expensive, however. The physical processes must be computed explicitly step by step in time.
[0035]
[0036] Such quasi-analytical approaches are extremely simplified, however. As a result, they tend to have large prediction errors. The predicted parameters, however, do inherently depend on the structural features of the network, and as such it should be possible to make a direct prediction from these features.
[0037]
[0038] In step 210, series of synthetic or real images are generated or collected bearing similar statistical and process parameter properties.
[0039] Such synthetic images can be generated using a suite of statistical or object based techniques. An example of such an object based technique can be one of the following: [0040] 1. Statistical reconstruction of three-dimensional porous media from two-dimensional images, by Anthony P. Roberts, Phys. Rev. E 56, 3203. 1997. [0041] 2. Prediction of permeability for porous media reconstructed using multiple-point statistics, Hiroshi Okabe and Martin J. Blunt, Phys. Rev. E 70, 066135, 2004. [0042] 3. 3D Stochastic Modelling of Heterogeneous Porous Media—Applications to Reservoir Rocks, Transport in Porous Media volume 65, pages 443-467(2006). [0043] 4. Comparing organic-hosted and intergranular pore networks: topography and topology in grains, gaps and bubbles, Matthew Andrew, Geological Society, London, Special Publications, 484, 3 Sep. 2018. [0044] 5. Reconstruction of three-dimensional porous media using generative adversarial neural networks, Lukas Mosser, Olivier Dubrule, and Martin J. Blunt, Phys. Rev. E 96, 043309, 2017.
[0045] In other examples, the images are not necessarily synthetic. For example, the real images could be images obtained from actual samples. In general, however, the synthetic images have the advantage that many can be created with relative ease.
[0046] The synthetic or hybrid or actual images are then segmented into individual phases, such as phases, pores, grains, or other structures in the images in step 212.
[0047] The process then bifurcates.
[0048] The process parameters and other physical properties are computed from these images using traditional modelling or physics simulation techniques in step 214. Examples of appropriate modeling tools include such open source projects as OpenFOAM, a large number of academic codes (e.g. https://www.imperial.ac.uk/earth-science/research/research-groups/perm/research/pore-scale-modelling/software/ or http://openpnm.org/), or internal industrial software projects (such as Schlumberger Direct HydroDynamic (DHD) simulation) as well as a suite of commercial software packages, including ThermoFisher Avizo Xlab, ThermoFisher eCore, ThermoFisher Pergeos, VolumeGraphics VGStudio, Comsol, Abaqus and Math2Market GeoDict simulation packages.
[0049] In parallel, the images are analyzed to separate physically touching objects in step 216. Objects are defined as contiguous regions belonging to a single phase. Approaches include morphological techniques or deep learning techniques.
[0050] Then in step 218, measurements of the objects are made.
[0051] Statistical feature vectors are constructed from the measurements in step 220. If separation/measurement/analysis is exclusively done in 2D, model is “2D.” If any or all is done in 3D, model is 3D.
[0052] In general, the extracted features could relate to a wide array of properties measured in the image, and may vary from predicted process (effective) parameter to predicted process (effective) parameter, and may be extracted in 2D or in 3D. Note that as it is possible to create a model from features extracted from a 2D slice from a 3D network, it is possible to create a predictive model which operates only from 2D data. This allows for effective multivariate predictions to be made even when the only data available is 2D.
[0053] A prediction model is then creates using multivariant regression of feature vectors to predicted physical property based on the constructed feature vectors and the computed physical properties in step 222.
[0054] Generally, statistics are computed across all the objects in the image. These statistics are then used in a multivariant description of the network as a whole. This multivariant description is then regressed against the parameter values using multivariant linear or non-linear regression to create the prediction model. In fact, any number of regression techniques may be used.
[0055] Then, in step 230, the prediction model is used to determine the physical properties of samples from 2D or 3D images. These samples are taken from mineral and/or oil/gas exploration and production. The samples could also come from manufacturing operations such as the manufacture of batteries and powder bed 3D additive manufacturing. The features vectors are extracted and given to the prediction model of step 222 to predict the parameters of the samples.
[0056]
[0057] Here, the images from steps 210 and 212 are fed into a convolutional neural network which maps from image area to single continuous variable.
[0058] Then, as before, a prediction model is created in step 222 using multivariant regression of feature vectors to predicted physical property based on the constructed feature vectors and the computed physical properties from step 214.
[0059]
[0060] In more detail, the N images 310 are simulated as described in connection with step 210 of
[0061] Another approach would be to perform a direct regression from the images using 2D or 3D networks (bypassing the need for structural statistical extraction).
[0062] Structural parameters can then be calculated 312 on real and/or synthetic (imaged) geometries as in step 218 and applied to predict the process parameter.
[0063] Incremental imaged and simulated volumes can be added to the training set for multivariant regression 314, or used as to create new regression sequences, or as the basis for transfer learning from the original set.
[0064]
[0065] The present approach could be extended to other machine learning tools such as neural networks. The model could be a trained neural network which takes as an input directly the 2D or 3D image structures from the training set, or the statistical feature set extracted from these images. Other alternatives are random forest regression or other multivariant regression models.
[0066] An example application of this technology is the prediction of permeability parameters from 2D images generated from light and/or electron microscopy.
[0067] Typically, such prediction of permeability is performed (at the pore scale) by performing 3D computational fluid dynamics on x-ray microscopy image datasets. The challenge with this approach is that the addressable spatial lengthscale with these (or, more precisely, its ratio with the voxel resolution of the volume) is extremely limited.
[0068] In contrast, the present approach can use multivariant regression based prediction, applied to 2D light and/or electron microscopy image data, which can be acquired over a much larger spatial lengthscale.
[0069] In general, sample permeability is critical to understand a range of applications from subsurface oil and gas flow, to carbon capture and storage, to filter performance, battery performance and more.
[0070] This takes the following process/workflow, which is described in connection with
[0071] First, a suite of segmented training images is required. These can be produced by a range of approaches as set forth in step 210: [0072] a. Image library. If samples have been imaged then these image datasets can be used for training, but a broad range of samples is required. In one example, the image datasets are 3D dataset from an x-ray microscope. This is shown in
[0075] Second, flow is simulated in each of these geometries. This is performed using standard computational fluid dynamics techniques using a physics simulation application 416.
[0076] Such a physics simulation application 416 is executed on a computer system 450 such as computer workstation or computer cluster or a cloud-based computer system. The computer system 450 has a hardware system 454 including one or more microprocessors and attendant memory, along with other storage resources. An operating system 452 will typically execute on the hardware system. The operating system 452 provides access to the compute resources for the various applications executing on the operation system such as the physics simulation applications 416.
[0077] In parallel, each geometry is analyzed to extract a statistical feature vector of the geometry in step 220. This uses image or 3D volume analysis app 418 that executes on the computer system 450 or another computer system. The 3D volume analysis app 418 performs following sub-workflow in one example:
a. 2D slices are randomly extracted from the volume. Each slice is analyzed independently.
b. Objects within each slice are separated as in step 216. This uses a multi-scale object separation routine. Pores and grains are separated independently. The multiscale object separation can be performed as follows: 1. a Euclidian (or chamfer) distance transform is calculated on the objects, 2. small object seeds are created by identifying local maxima in the distance transform, 3. large object seeds are created by segmenting the distance transform, 4. the union between the two seed objects is evaluated, 5. this image is then analyzed to identify connected components, with each forming the seed of a separated object, and 6. These seeds are then grown on the landscape of the distance transform image using a watershed algorithm.
c. This generates a separated pore and grain image. A final “pore throat” image is created by 3D volume analysis app 418 evaluating the regions when two separated pore objects touch.
d. Measurements are then made on each of these objects, creating a set of features as in step 218 by the 3D volume analysis app 418. Many different measurements are possible. Example measurements include:
TABLE-US-00001 TABLE 1 Feature/unit Inscribed circle radius/μm Area/μm.sup.2 Euler number Coordination number Perimeter/μm Convexity
e. From this list of features, statistical measures are made on each feature as in step 220. A range of different statistical measures are possible, but an example evaluated for permeability prediction from thin section are:
TABLE-US-00002 TABLE 2 Feature/unit Throat inscribed radius standard deviation/μm Pore inscribed radius mean/μm Average throat inscribed radius/μm Largest throat inscribed radius/μm Pore inscribed radius standard deviation/μm Largest pore inscribed radius/μm Pore inscribed radius area weighted mean/μm Porosity Fraction of pores with coordination number 2 Pixel size/μm Grain inscribed radius standard deviation/μm Coordination number 0 Grain inscribed radius mean/μm Area weighted mean pore area/μm.sup.2 Largest grain inscribed radius/μm.sup.2 Lognorm Fitted Pore Area/μm.sup.2 Pore area standard deviation/μm.sup.2 Largest pore area/μm.sup.2 Grain inscribed radius weighted mean/μm Fraction of pores with coordination number 3 Mean pore area/μm.sup.2 Maximum pore area/μm.sup.2 Fraction of pores with coordination number 4 Fraction of pores with coordination number 1 Euler number (porosity) Largest grain area/μm.sup.2 Fraction of pores with coordination number 5 Lognorm fitted grain area/μm.sup.2 Euler number (volume weighted, grain) Euler number (grain) Lognorm fitted pore number standard deviation Euler number (volume weighted, pore) Fraction of pores with coordination number 6 Fraction of pores with coordination number 7
f. This then provides a statistical feature vector description of the pore network. Measurements (from step d) can be aggregated across multiple slices prior to statistical measurement to increase the robustness of the measure.
[0078] Third, a prediction model 426 is created using multivariant regression, e.g. using a Random Forest, multi-layer perceptron, support vector machine or other machine learning approach as in step 222. In the illustrated example, the machine learning is performed by a machine learning app 420 executing on the computer system 450 which received that statistical feature vector descriptions from the geometry app 418.
[0079] Finally, once this model 426 is created, it can be applied on any 2D image dataset (e.g. image dataset from a light microscope or electron microscope), once that data has been segmented into pore and grain (features were only extracted from 2D slices) as in step 230. This allows for much larger spatial lengthscales to be applied in a reasonable period of time.
[0080] In the illustrated example, new samples 412 produced by the manufacturing or execution operation 410 are image in a SEM 422 and provided directly to the machine learning app 420 that applies the prediction model 426 or first provided to the geometry app that then provides their statistical feature vector descriptions to the machine learning app 420
[0081] While this invention has been particularly shown and described with references to preferred embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the scope of the invention encompassed by the appended claims.