Systems and methods for blind multi-spectral image fusion
11481876 · 2022-10-25
Assignee
Inventors
- Dehong Liu (Lexington, MA)
- Lantao Yu (Houston, TX, US)
- Yanting Ma (Allston, MA, US)
- Hassan Mansour (Boston, MA)
- Petros Boufounos (Winchester, MA)
Cpc classification
G06T3/4076
PHYSICS
International classification
G06T3/40
PHYSICS
Abstract
Systems, methods and apparatus for image processing for reconstructing a super resolution image from multispectral (MS) images. Receive image data and initialize a fused image using a panchromatic (PAN) image, and estimate a blur kernel between the PAN image and the MS images as an initialization function. Iteratively, fuse a MS image with an associated PAN image of a scene using a fusing algorithm. Each iteration includes: update the blur kernel based on a Second-Order Total Generalized Variation function to regularize a kernel shape; fuse the PAN image and MS images with the updated blur kernel based on a local Laplacian prior function to regularize the high-resolution information to obtain an estimated fused image; compute a relative error between the estimated fused image of the current iteration and a previous estimated fused image from a previous iteration, to a predetermined threshold, to stop iterations stop, to obtain a PAN-sharpened image.
Claims
1. A system for reconstructing a super resolution image from multispectral (MS) images, having a transceiver to accept data, a memory to store the data, wherein the data including MS images and a panchromatic (PAN) of a scene, each MS image is associated with the PAN image, as well as a processing device operatively connected to the transceiver and the memory, the system comprising that the processing device is configured to: initialize a fused image using the PAN image, and estimate a blur kernel between the PAN image and the MS images as an initialization function; iteratively, fuse a MS image with an associated PAN image of the scene using a fusing algorithm by a processor, each iteration includes: update the blur kernel based on a Second-Order Total Generalized Variation (TGV.sup.2) function to regularize a kernel shape; fuse the PAN image and MS images with the updated blur kernel based on a local Laplacian prior (LLP) function to regularize the high-resolution similarity between the PAN image and the fused MS image to obtain an estimated fused image; compute a relative error between the estimated fused image of the current iteration and a previous estimated fused image from a previous iteration, wherein, when the relative error is less than a predetermined threshold, the iterations stop, resulting in obtaining a PAN-sharpened image; and output, via an output interface in communication with the processor, the PAN-sharpened image to a communication network or to a display device.
2. The system of claim 1, wherein the PAN image used to initialize the fused image is a ridge PAN image.
3. The system of claim 1, wherein the blur kernel is a rigid transformation blur kernel, and the initialization function is an initial blur kernel function.
4. The system of claim 1, wherein the LLP regularizes a relationship between high-frequency components of the MS images and PAN image, yielding a level of a fusion performance that is greater than a level of a fusion performance using local gradient constraints.
5. The system of claim 4, wherein the LLP is a second-order gradient, such that the LLP is generalized to a second gradient or a higher order gradient.
6. The system of claim 1, wherein the TGV.sup.2 function is operable when an assumption that an image is piecewise constant is not valid in reconstructing images, such that the piecewise constant images are captured using the TGV.sup.2 function during the image reconstruction.
7. The system of claim 6, wherein the TGV2 is a regularizer on the blur kernel, which is assumed to be smooth and centralized according to the TGV2.
8. The system of claim 1, wherein the MS images are obtained from a MS image sensor having a color filter array and positioned at a first optical axis and the PAN images are obtained from a PAN image sensor positioned at a second optical axis that converges at an angle with the first optical axis.
9. An apparatus having computer storage including a computer-readable storage medium, and a hardware processor device operatively coupled to the computer storage and to reconstruct spatial resolution of an image of a scene captured within multi-spectral (MS) images and panchromatic (PAN) images, the MS images obtained from a MS image sensor having a color filter array and positioned at a first optical axis, and the PAN images obtained from a PAN image sensor positioned at a second optical axis that is substantially parallel to the first optical axis, wherein, to reconstruct the spatial resolution of the image, the apparatus comprising that the hardware processor device is to: initialize a fused image using a PAN image, and estimate a blur kernel between the PAN image and the MS images using an initialization function; iteratively, fuse a MS image with an associated PAN image of the scene using a fusing algorithm by a processor, each iteration includes: update the blur kernel based on a Second-Order Total Generalized Variation (TGV.sup.2) function to regularize a kernel shape; fuse the PAN image and MS images with the updated blur kernel based on a local Laplacian prior (LLP) function to regularize the high-resolution similarity between the PAN and the fused MS image to obtain an estimated fused image; compute a relative error between the estimated fused image of the current iteration and a previous estimated fused image from a previous iteration, wherein, when the relative error is less than a predetermined threshold, the iterations stop, resulting in obtaining a PAN-sharpened image; and output, via an output interface in communication with the processor, the PAN-sharpened image to a communication network or to a display device.
10. The apparatus of claim 9, wherein the MS images are low resolution images and are obtained from the MS image sensor optically coupled to a first imaging lens, and the PAN images are high resolution images and are obtained from the PAN image sensor, the MS image sensor and the PAN image sensor have substantially identical focal plane arrays of substantially identical photosensitive elements, and wherein the MS image sensor and the PAN image sensor are set in substantially a single geometric plane such that the focal plane arrays receive optical projections of substantially an identical version of the scene.
11. The apparatus of claim 9, wherein the MS images are captured at a first frame rate and the PAN images are captured at a second frame rate different than or the same as the first frame rate.
12. The apparatus of claim 9, wherein the blur kernel combines a Point Spread Function (PSF) function and a shift such as a rigid transformation, together.
13. The system of claim 9, wherein the MS images are obtained from a MS image sensor having a color filter array and positioned at a first optical axis and the PAN images are obtained from a PAN image sensor positioned at a second optical axis that converges at an angle with the first optical axis.
14. A system for reconstructing a super resolution image from multispectral (MS) images, and an input interface to accept data, along with a memory to store the data, the data including MS images and panchromatic (PAN) images of a scene, each MS image is associated with a PAN image, such that a hardware processing device is operatively connected to the input interface and the memory, the system comprising that the hardware processing device is configured to: initialize a fused image using a rigid PAN image; estimate a rigid transformation blur kernel between the PAN image and the MS images as an initial blur kernel function; iteratively, fuse a MS image with an associated PAN image of the scene using a fusing algorithm by a processor, each iteration includes: update the blur kernel based on a Second-Order Total Generalized Variation (TGV.sup.2) function to regularize a kernel shape; fuse the PAN image and MS images with the updated blur kernel based on a local Laplacian prior (LLP) function to regularize the high-resolution similarity between the PAN image and the fused MS image to obtain an estimated fused image; compute a relative error between the estimated fused image of the current iteration and a previous estimated fused image from a previous iteration, wherein, when the relative error is less than a predetermined threshold, the iterations stop, resulting in obtaining a PAN-sharpened image; and output, via an output interface in communication with the processor, the PAN-sharpened image to a communication network or to a display device.
15. The system of claim 14, wherein the MS images are obtained from a MS image sensor having a color filter array and positioned at a first optical axis and the PAN images are obtained from a PAN image sensor positioned at a second optical axis that converges at an angle with the first optical axis.
16. The system of claim 14, wherein the data accepted by the input interface includes some data obtained from sensors including at least one MS image sensor device and at least one PAN image sensor device.
17. The system of claim 14, wherein the PAN image used to initialize the fused image is a ridge PAN image.
18. The system of claim 14, wherein the blur kernel is a rigid transformation blur kernel, and the initialization function is an initial blur kernel function.
19. A non-transitory machine-readable medium including instructions stored thereon which, when executed by processing circuitry, configure the processing circuitry to perform operations to sharpen a multi-spectral (MS) image using data from a panchromatic (PAN) image, the operations for: receiving data, the data including MS images and a panchromatic (PAN) image of a scene, each MS image is associated with the PAN image; initializing a fused image using the PAN image, and estimate a blur kernel between the PAN image and the MS images to obtain a blur kernel using an initialization function; iteratively, fuse a MS image with an associated PAN image of the scene using a fusing algorithm by a processor, each iteration includes: updating the blur kernel based on a Second-Order Total Generalized Variation (TGV.sup.2) function to regularize a kernel shape; fusing the PAN image and MS images with the updated blur kernel based on a local Laplacian prior (LLP) function to regularize the high-resolution similarity between the PAN image and the fused MS image to obtain an estimated fused image; computing a relative error between the estimated fused image of the current iteration and a previous estimated fused image from a previous iteration, wherein, when the relative error is less than a predetermined threshold, the iterations stop, resulting in obtaining a PAN-sharpened image; and outputting the PAN-sharpened image to a communication network or to a display device via an output interface in communication with the processor.
20. The non-transitory machine-readable medium of claim 19, further including instructions stored thereon which, when executed by a machine, are configured for the machine to perform operations to create a PAN image with about a same resolution as a resolution of a MS image by down-sampling PAN image data stored in the memory, or determining PAN image data from the MS image data, such that the received data, received via a transceiver device in communication with the non-transitory machine-readable medium and processing circuitry, includes some data obtained from sensors including at least one MS image sensor device and at least one PAN image sensor device.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) The presently disclosed embodiments will be further explained with reference to the attached drawings. The drawings shown are not necessarily to scale, with emphasis instead generally being placed upon illustrating the principles of the presently disclosed embodiments.
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12) While the above-identified drawings set forth presently disclosed embodiments, other embodiments are also contemplated, as noted in the discussion. This disclosure presents illustrative embodiments by way of representation and not limitation. Numerous other modifications and embodiments can be devised by those skilled in the art which fall within the scope and spirit of the principles of the presently disclosed embodiments.
DETAILED DESCRIPTION
(13)
(14) Step 110 of method 100A can include receiving data, the data including MS images and a panchromatic (PAN) image of a scene, each MS image is associated with the PAN image.
(15) Further, each MS image includes multiple channels, each channel is associated with a frequency band, such that an image of a channel represents the frequency response within the associated frequency band. It is possible the data can be stored in a memory. For example, the data can be stored in one or more databases of a computer readable memory, such that the processor or hardware processor is in communication with the computer readable memory and the input interface or a transceiver.
(16) Step 115 of
(17) Step 120 of
(18) Step 125 of
(19) Step 130 of
(20) Step 135 of
(21) Step 140 of
(22) Some methods of the present disclosure use a blind multi-spectral (MS) Image fusion method using local Laplacian prior (LLP) and second-order total generalized variation (TGV.sup.2). The LLP regularizes the relationship between high-frequency components of MS and PAN images, yielding better fusion performance than local gradient constraints. Wherein, the TGV.sup.2 regularizes the blur kernel with more robustness to noise and more accurate estimation of the blur kernel than other existing sparsity-driven priors. From experimentation, results exhibited consistent better performance on fusing mis-registered MS and panchromatic images than the conventional state-of-the-art methods in terms of visual quality and multiple quantitative metrics. Further, as exhibited from experimentation, the methods of the present disclosure achieved a fast convergence in a short computational time with a warm start, than the conventional state-of-the-art methods. Wherein the blind fusion algorithm outperformed conventional deep-learning based methods in regions with abundant edges and textures such as Cuprite, Moffett, and Los Angeles images, and is comparable in regions without too much edges such as Cambria Fire image.
(23) Also, some aspects of the embodiments of the present disclosure include novel methods for misaligned MS image pan-sharpening based on the local Laplacian prior (LLP) and the Second-Order Total Generalized Variation (TGV2). Numerical experiments show that some method approaches significantly outperform conventional optimization-based and deep learning-based baseline approaches. Moreover, some embodiments of the present disclosure have a better generalization ability than conventional deep learning-based methods, due in part to not having or without external training data, and thus, provide substantial flexibility and adaptability to deal with multi-spectral imagery from a large variety of imaging platforms.
(24)
(25) The signal data can include multi-spectral (MS) image data gathered by at least one external sensor 14 and acquired by the input interface 13 or from an external memory device 15, or some other means of communication either wired or wireless. For example, the signal data can be acquired by the processor 12 either directly or indirectly, e.g., a memory transfer device, or a wireless communication like device. It is possible, a user interface 17 having a keyboard (not shown) can be in communication with the processor 12 and a computer readable memory, and can acquire and store the MS and PAN images in the computer readable memory 10 and other data, upon receiving an input from a surface of the keyboard of the user interface 17 by a user.
(26) Still referring to
(27)
(28)
(29)
(30) Step 1, 301 of
(31) Step 2, 303 of
(32) Step 3, 305 to step 11, 321 of
(33)
in which the first component is the data fidelity term, u∈R.sup.n.sup.
(34)
where parameters are defined as follows: λ is a scalar factor; ω.sub.j is the j.sup.th square window of size (2r+1)(2r+1) in an H×W image, with r an integer; k refers to the k.sup.th element within the window, k=1, 2, . . . , (2r+1).sup.2; a.sub.i,j and c.sub.i,j are both constant coefficients of the linear affine transform in window ω.sub.1, corresponding to the i.sup.th band; Z.sub.i is the i.sup.th band of Z, L(.Math.) is a function that computes the Laplacian of the input image, i.e., L(Z)=Z.Math.S, with
(35)
(36) Step 3, 305 of
(37)
(38) Step 5, 309 and step 6, 311 of
(39)
Where ∇u=[∇.sub.hu ∇.sub.vu]∈R.sup.n.sup.
(40)
∥X∥.sub.2,1=Σ.sub.i=1.sup.n√{square root over (Σ.sub.j=1.sup.mx.sub.i,j.sup.2)}, and α.sub.1, α.sub.2 are both scalars that control the regularization strength of p's approximation to ∇u and of the partial derivatives of p. S={S∈R.sup.n.sup.
(41) Step 5, 309 of
(42)
with μ.sub.1, μ.sub.2, μ.sub.3>0. We solve the problem using the alternating direction method of multipliers (ADMM) by alternating between a succession of minimization steps and update steps.
(43) The minimization subproblems of x and y are similar to each other and the solutions are given by component-wise soft-thresholding.
(44) Step 5, 309 of
(45)
(46) Step 5, 309 of
(47)
using conjugate gradient descent and then projects the solution onto the simplex S.
(48) Step 6, 311 of
(49)
(50) The problem can be solved efficiently by making use of the fast Fourier transform.
(51) Step 8, 315 of
(52)
similar to guided image filtering, a.sub.i,j and c.sub.i,j can be stably computed using L(z.sub.i)'s local window as the input image and 's local window as the guide image.
(53) Step 9, 317 of
(54) The Z-subproblem in each individual channel is reformulated as
(55)
(56) Equation (12) has a closed-form solution:
Z.sub.i=(B.sup.TD.sup.TDB+λL.sup.TL).sup.−1(B.sup.TD.sup.TX.sub.i+λ{circumflex over (L)}.sub.i.sup.x) (14).
(57) Similarly, we use the Fast Fourier Transform to accelerate the computation since B is a Toeplitz matrix.
(58) Step 10, 319 of
(59)
until the relative error between the estimated fused image of the current iteration
(60)
wherein ϵ is a pre-defined threshold.
(61) Step 11, 321 of
(62)
(63) In comparison, BHMIFGLR failed to generate fused MS images with consistent performance; the blur estimated kernel estimated, shown in
(64)
(65)
(66)
(67) Features
(68) A system for reconstructing a super resolution image from multispectral (MS) images. The system having a transceiver to accept data. A memory to store the data, the data including MS images and a panchromatic (PAN) of a scene, each MS image is associated with the PAN image. A processing device operatively connected to the transceiver and the memory. The system comprising that the processing device is configured to initialize a fused image using the PAN image, and estimate a blur kernel between the PAN image and the MS images as an initialization function. Iteratively, fuse a MS image with an associated PAN image of the scene using a fusing algorithm by a processor. Each iteration includes: update the blur kernel based on a Second-Order Total Generalized Variation (TGV.sup.2) function to regularize a kernel shape; fuse the PAN image and MS images with the updated blur kernel based on a local Laplacian prior (LLP) function to regularize the high-resolution similarity between the PAN image and the fused MS image to obtain an estimated fused image; compute a relative error between the estimated fused image of the current iteration and a previous estimated fused image from a previous iteration, wherein, when the relative error is less than a predetermined threshold, the iterations stop, resulting in obtaining a PAN-sharpened image. Output, via an output interface in communication with the processor, the PAN-sharpened image to a communication network or to a display device. Wherein, it is contemplated is that the system, can include any combination of the different aspects listed below, regarding the above system. In particular, the following aspects are intended to either individually or in combination, create one or more embodiments based on the one or more combination of aspects listed below, for the above recited system.
(69) An aspect is that the PAN image used to initialize the fused image is a ridge PAN image. Another aspect is that the blur kernel is a rigid transformation blur kernel, and the initialization function is an initial blur kernel function.
(70) Wherein an aspect can include that the LLP regularizes a relationship between high-frequency components of the MS images and PAN image, yielding a level of a fusion performance that is greater than a level of a fusion performance using local gradient constraints. Wherein an aspect is that the LLP is a second-order gradient, such that the LLP is generalized to a second gradient or a higher order gradient.
(71) Another aspect includes the TGV.sup.2 function is operable when an assumption that an image is piecewise constant is not valid in reconstructing images, such that the piecewise constant images are captured using the TGV.sup.2 function during the image reconstruction. Wherein an aspect includes the TGV2 is a regularizer on the blur kernel, which is assumed to be smooth and centralized according to the TGV2.
(72) Another aspect is that the MS images are obtained from a MS image sensor having a color filter array and positioned at a first optical axis and the PAN images are obtained from a PAN image sensor positioned at a second optical axis that converges at an angle with the first optical axis.
(73) An aspect is that the MS images are low resolution images and are obtained from the MS image sensor optically coupled to a first imaging lens, and the PAN images are high resolution images and are obtained from the PAN image sensor, the MS image sensor and the PAN image sensor have substantially identical focal plane arrays of substantially identical photosensitive elements, and wherein the MS image sensor and the PAN image sensor are set in substantially a single geometric plane such that the focal plane arrays receive optical projections of substantially an identical version of the scene.
(74) Another aspect is the MS images are captured at a first frame rate and the PAN images are captured at a second frame rate different than or the same as the first frame rate. Wherein an aspect can include the blur kernel combines a Point Spread Function (PSF) function and a shift such as a rigid transformation, together. It is possible another aspect is that the MS images are obtained from a MS image sensor having a color filter array and positioned at a first optical axis and the PAN images are obtained from a PAN image sensor positioned at a second optical axis that converges at an angle with the first optical axis.
(75) An aspect is that the MS images are obtained from a MS image sensor having a color filter array and positioned at a first optical axis and the PAN images are obtained from a PAN image sensor positioned at a second optical axis that converges at an angle with the first optical axis. Further, contemplated is that an aspect is the data accepted by the input interface includes some data obtained from sensors including at least one MS image sensor device and at least one PAN image sensor device. Another aspect is that the PAN image used to initialize the fused image is a ridge PAN image. It is possible that an aspect is that the blur kernel is a rigid transformation blur kernel, and the initialization function is an initial blur kernel function.
(76) Wherein, an aspect can further include instructions stored thereon which, when executed by a machine, are configured for the machine to perform operations to create a PAN image with about a same resolution as a resolution of a MS image by down-sampling PAN image data stored in the memory, or determining PAN image data from the MS image data, such that the received data, received via a transceiver device in communication with the non-transitory machine-readable medium and processing circuitry, includes some data obtained from sensors including at least one MS image sensor device and at least one PAN image sensor device. An aspect may be that the blur kernel jointly combines a Point Spread Function and a rigid transformation blur kernel.
DEFINITIONS
(77) According to aspects of the present disclosure, and based on experimentation, the following definitions have been established, and certainly are not a complete definition of each phrase or term. Wherein the provided definitions are merely provided as an example, based upon learnings from experimentation, wherein other interpretations, definitions, and other aspects may pertain. However, for at least a mere basic preview of the phrase or term presented, such definitions have been provided. Further, the definitions below cannot be viewed as prior art since the knowledge gained is from experimentation only.
(78) Blind Deconvolution: Blind deconvolution is a deconvolution technique that permits recovery of the target scene from a single or set of “blurred” images in the presence of a poorly determined or unknown point spread function (PSF). (Note: In this patent, the unknown blur kernel is basically a rigid transformed PSF.) Regular linear and non-linear deconvolution techniques utilize a known PSF. For blind deconvolution, the PSF is estimated from the image or image set, allowing the deconvolution to be performed. Blind deconvolution can be performed iteratively, whereby each iteration improves the estimation of the PSF and the scene, or non-iteratively, where one application of the algorithm, based on exterior information, extracts the PSF. Iterative methods include maximum a posteriori estimation and expectation-maximization algorithms. A good estimate of the PSF is helpful for quicker convergence but not necessary.
(79) Some challenges of Blind deconvolution can be that both input image and blur kernel must live in fixed subspace. That means input image, represented by w, has to be written as w=Bh, where B is random matrix of size L by K (K<L) and h is of size K×1, whereas blur kernel, if represented by x, has to be written as x=Cm, where C is random matrix of size L by N (N<L) and in is of size N×1. Observed image, if represented by y, given by y=w*x, can only be reconstructed if L>=K+N.
(80) Point Spread Function (PSF): PSF describes the response of an imaging system to a point source or point object. A more general term for the PSF is a system's impulse response, the PSF being the impulse response of a focused optical system. The PSF in many contexts can be thought of as the extended blob in an image that represents a single point object. In functional terms it is the spatial domain version of the optical transfer function of the imaging system. It is a useful concept in Fourier optics, astronomical imaging, medical imaging, electron microscopy and other imaging techniques such as 3D microscopy (like in confocal laser scanning microscopy) and fluorescence microscopy. The degree of spreading (blurring) of the point object is a measure for the quality of an imaging system. In non-coherent imaging systems such as fluorescent microscopes, telescopes or optical microscopes, the image formation process is linear in the image intensity and described by linear system theory. This means that when two objects A and B are imaged simultaneously, the resulting image is equal to the sum of the independently imaged objects. In other words: the imaging of A is unaffected by the imaging of B and vice versa, owing to the non-interacting property of photons. In space-invariant system, i.e. the PSF is the same everywhere in the imaging space, the image of a complex object is then the convolution of the true object and the PSF. However, when the detected light is coherent, image formation is linear in the complex field. The recorded intensity image then can show cancellations or other non-linear effects.
(81) Deep Image Prior: Deep image prior is a type of convolutional neural network used to enhance a given image with no prior training data other than the image itself. A neural-network is randomly initialized and used as prior to solve inverse problems such as noise reduction, super-resolution, and inpainting. Image statistics is captured by the structure of a convolutional image generator rather than by any previously learned capabilities.
(82) Resolution tradeoffs using some Sensors: Some aspects learned from experimentation includes that all sensors can have a fixed signal-to-noise ratio that can be a function of the hardware design. The energy reflected by the target needs to have a signal level large enough for the target to be detected by the sensor. The signal level of the reflected energy increases if the signal is collected over a larger instantaneous field of view (IFOV) or if it is collected over a broader spectral bandwidth. Collecting energy over a larger IFOV reduces the spatial resolution while collecting it over a larger bandwidth reduces its spectral resolution. Thus, there is a tradeoff between the spatial and spectral resolutions of the sensor. As noted above, a high spatial resolution can accurately discern small or narrow features like roads, automobiles, etc. A high spectral resolution allows the detection of minor spectral changes, like those due to vegetation stress or molecular absorption. It seemed that most optical remote sensing satellites carry two types of sensors—the panchromatic and the multispectral sensors. The multispectral sensor records signals in narrow bands over a wide IFOV while the panchromatic sensor records signals over a narrower IFOV and over a broad range of the spectrum. Thus, the multispectral (MS) bands have a higher spectral resolution, but a lower spatial resolution compared to the associated panchromatic (PAN) band, which has a higher spatial resolution and a lower spectral resolution.
(83) Alternating Direction Method of Multipliers (ADMM): ADMM is a variant of the augmented Lagrangian scheme that uses partial updates for the dual variables. This method is often applied to solve problems such as
(84)
This is equivalent to a constrained problem
(85)
subjet to x=y. Though this change may seem trivial, the problem can now be attacked using methods of constrained optimization (in particular, the augmented Lagrangian method), and the objective function is separable in x and y. The dual update requires solving a proximity function in x and y at the same time; the ADMM technique allows this problem to be solved approximately by first solving for x with y fixed, and then solving for y with x fixed. Rather than iterate until convergence (like the Jacobi method), the algorithm proceeds directly to updating the dual variable and then repeating the process. This is not equivalent to the exact minimization, but surprisingly, it can still be shown that this method converges to the right answer (under some assumptions). Because of this approximation, the algorithm is distinct from the pure augmented Lagrangian method.
(86) Total Variation (TV) and Total Generalized Variation (TGV): TV based strategies, can include regularization for parallel imaging, such as in iterative reconstruction of under-sampled image data sets. TV models can have a benefit that they are well suited to remove random noise, while preserving edges in the image. However, an assumption of TV is that the images consist of regions, which are piecewise constant. What was learned is that the use of TV can often lead to staircasing artifacts and result in patchy, sketch type images which appear unnatural. However, using TGV which may be equivalent to TV in terms of edge preservation and noise removal, TGV can also be applied in imaging situations where an assumption that the image is piecewise constant is not valid. As a result, an application of TGV in imaging can be less restrictive. For example, TGV can be applied for image denoising and during iterative image reconstruction of under-sampled image data sets, was found to possibly yielding results that are superior to conventional TV. Where TGV may be capable to measure, in some sense, image characteristics up to a certain order of differentiation. At least one aspect noted is that TV only takes the first derivative into account. TGV is a semi-norm of a Banach space, associated variational problems that fit well into a well-developed mathematical theory of convex optimization problems, especially with respect to analysis and computational realization. Moreover, each function of bounded variation admits a finite TGV value, making the notion suitable for images. Which means that piecewise constant images can be captured with the TGV model which even extends the TV model. Finally, TGV is translation invariant as well as rotationally invariant, meaning that it is in conformance with the requirement that images are measured independent from the actual viewpoint. However, it was learned that using TGV.sup.2 as a regularizer can lead to an absence of the staircasing effect which is often observed in TV regularization.
(87) Piecewise Constant Function: A function is said to be piecewise constant if it is locally constant in connected regions separated by a possibly infinite number of lower-dimensional boundaries. The Heaviside step function, rectangle function, and square wave are examples of one-dimensional piecewise constant functions. In mathematics, a piecewise-defined function (also called a piecewise function, a hybrid function, or definition by cases) is a function defined by multiple sub-functions, each sub-function applying to a certain interval of the main function's domain, a sub-domain. Piecewise is actually a way of expressing the function, rather than a characteristic of the function itself, but with additional qualification, it can describe the nature of the function. For example, a piecewise polynomial function is a function that is a polynomial on each of its sub-domains, but possibly a different one on each. The word piecewise is also used to describe any property of a piecewise-defined function that holds for each piece but not necessarily hold for the whole domain of the function. A function is piecewise differentiable or piecewise continuously differentiable if each piece is differentiable throughout its subdomain, even though the whole function may not be differentiable at the points between the pieces. In convex analysis, the notion of a derivative may be replaced by that of the subderivative for piecewise functions. Although the “pieces” in a piecewise definition need not be intervals, a function is not called “piecewise linear” or “piecewise continuous” or “piecewise differentiable” unless the pieces are intervals.
(88) Actual viewpoint: A viewpoint refers to the position we take the photograph from. This will also be the position you place the viewer in when they are looking at your finished shot. The viewpoint can dramatically change the feel of the photograph. A transformation of particular interest is viewpoint (i.e., camera panning, zooming, and translation). Cast as an image transformation, a change in camera viewpoint can be modeled as a mapping, or warp, between pixels in one or more basis views and pixels in a new image, representing a synthetic view of the same scene. Learned from experimentation is that there are some factors to consider in addressing an actual viewpoint, such as: measurability: Sufficient information to compute the transformation must be automatically or semi-automatically extracted from the basis images; correctness: Each synthesized image should be physically correct, i.e., it should correspond to what the real scene would look like as a result of the specified scene transformation; and synthesis: New algorithms must be developed for image-based scene transformations. The techniques should be robust, easy to use, and general enough to handle complex real-world objects and scenes.
(89) An image moment is a certain particular weighted average (moment) of the image pixels' intensities, or a function of such moments, usually chosen to have some attractive property or interpretation.
(90) Moments invariant: Moments are well-known for their application in image analysis, since they can be used to derive invariants with respect to specific transformation classes. The term invariant moments is often abused in this context. However, while moment invariants are invariants that are formed from moments, the only moments that are invariants themselves are the central moments. Note that the invariants detailed are exactly invariant only in the continuous domain. In a discrete domain, neither scaling nor rotation are well defined: a discrete image transformed in such a way is generally an approximation, and the transformation is not reversible. These invariants therefore are only approximately invariant when describing a shape in a discrete image. Translation invariants: The central moments μ.sub.i j of any order are, by construction, invariant with respect to translations, i.e. in Euclidean geometry, a translation is a geometric transformation that moves every point of a figure or a space by the same distance in a given direction.
(91)
(92) The computer 711 can include a power source 754, depending upon the application the power source 754 may be optionally located outside of the computer 711. Linked through bus 756 can be a user input interface 757 adapted to connect to a display device 648, wherein the display device 748 can include a computer monitor, camera, television, projector, or mobile device, among others. A printer interface 759 can also be connected through bus 756 and adapted to connect to a printing device 732, wherein the printing device 732 can include a liquid inkjet printer, solid ink printer, large-scale commercial printer, thermal printer, UV printer, or dye-sublimation printer, among others. A network interface controller (NIC) 734 is adapted to connect through the bus 756 to a network 736, wherein image data or other data, among other things, can be rendered on a third-party display device, third party imaging device, and/or third-party printing device outside of the computer 711. The computer/processor 711 can include a GPS 701 connected to bus 756. Further,
(93) Still referring to
(94) The description provides exemplary embodiments only, and is not intended to limit the scope, applicability, or configuration of the disclosure. Rather, the following description of the exemplary embodiments will provide those skilled in the art with an enabling description for implementing one or more exemplary embodiments. Contemplated are various changes that may be made in the function and arrangement of elements without departing from the spirit and scope of the subject matter disclosed as set forth in the appended claims.
(95) Specific details are given in the following description to provide a thorough understanding of the embodiments. However, understood by one of ordinary skill in the art can be that the embodiments may be practiced without these specific details. For example, systems, processes, and other elements in the subject matter disclosed may be shown as components in block diagram form in order not to obscure the embodiments in unnecessary detail. In other instances, well-known processes, structures, and techniques may be shown without unnecessary detail in order to avoid obscuring the embodiments. Further, like reference numbers and designations in the various drawings indicated like elements.
(96) Also, individual embodiments may be described as a process which is depicted as a flowchart, a flow diagram, a data flow diagram, a structure diagram, or a block diagram. Although a flowchart may describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently. In addition, the order of the operations may be re-arranged. A process may be terminated when its operations are completed, but may have additional steps not discussed or included in a figure. Furthermore, not all operations in any particularly described process may occur in all embodiments. A process may correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a function, the function's termination can correspond to a return of the function to the calling function or the main function.
(97) Furthermore, embodiments of the subject matter disclosed may be implemented, at least in part, either manually or automatically. Manual or automatic implementations may be executed, or at least assisted, through the use of machines, hardware, software, firmware, middleware, microcode, hardware description languages, or any combination thereof. When implemented in software, firmware, middleware or microcode, the program code or code segments to perform the necessary tasks may be stored in a machine readable medium. A processor(s) may perform the necessary tasks.
(98) The above-described embodiments of the present disclosure can be implemented in any of numerous ways. For example, the embodiments may be implemented using hardware, software or a combination thereof. When implemented in software, the software code can be executed on any suitable processor or collection of processors, whether provided in a single computer or distributed among multiple computers.
(99) Also, the various methods or processes outlined herein may be coded as software that is executable on one or more processors that employ any one of a variety of operating systems or platforms. Additionally, such software may be written using any of a number of suitable programming languages and/or programming or scripting tools, and also may be compiled as executable machine language code or intermediate code that is executed on a framework or virtual machine. Typically, the functionality of the program modules may be combined or distributed as desired in various embodiments.
(100) Also, the embodiments of the present disclosure may be embodied as a method, of which an example has been provided. The acts performed as part of the method may be ordered in any suitable way. Accordingly, embodiments may be constructed in which acts are performed in an order different than illustrated, which may include performing some acts concurrently, even though shown as sequential acts in illustrative embodiments. Further, use of ordinal terms such as first, second, in the claims to modify a claim element does not by itself connote any priority, precedence, or order of one claim element over another or the temporal order in which acts of a method are performed, but are used merely as labels to distinguish one claim element having a certain name from another element having a same name (but for use of the ordinal term) to distinguish the claim elements.
(101) Although the present disclosure has been described with reference to certain preferred embodiments, it is to be understood that various other adaptations and modifications can be made within the spirit and scope of the present disclosure. Therefore, it is the aspect of the append claims to cover all such variations and modifications as come within the true spirit and scope of the present disclosure.