HIGHLY-SCALABLE IMAGE RECONSTRUCTION USING DEEP CONVOLUTIONAL NEURAL NETWORKS WITH BANDPASS FILTERING
20190257905 ยท 2019-08-22
Inventors
Cpc classification
G01R33/5611
PHYSICS
G01R33/5608
PHYSICS
G01R33/56545
PHYSICS
G01R33/56509
PHYSICS
International classification
G01R33/56
PHYSICS
G01R33/561
PHYSICS
G01R33/565
PHYSICS
Abstract
A method for magnetic resonance imaging (MRI) scans a field of view and acquires sub-sampled multi-channel k-space data U. An imaging model A is estimated. Sub-sampled multi-channel k-space data U is divided into sub-sampled k-space patches, each of which is processed using a deep convolutional neural network (ConvNet) to produce corresponding fully-sampled k-space patches, which are assembled to form fully-sampled k-space data V, which is transformed to image space using the imaging model adjoint A.sub.adj to produce an image domain MRI image. The processing of each k-space patch u.sub.i preferably includes applying the k-space patch u.sub.i as input to the ConvNet to infer an image space bandpass-filtered image y.sub.i, where the ConvNet comprises repeated de-noising blocks and data-consistency blocks; and estimating the fully-sampled k-space patch v.sub.i from the image space bandpass-filtered image y.sub.i using the imaging model A and a mask matrix.
Claims
1. A method for magnetic resonance imaging (MRI) comprising: (a) scanning a field of view using an MRI apparatus; (b) acquiring sub-sampled multi-channel k-space data U representative of MRI signals in the field of view; (c) estimating an imaging model A and corresponding model adjoint A.sub.adj by estimating a sensitivity profile map; (d) dividing sub-sampled multi-channel k-space data U into sub-sampled k-space patches; (e) processing the sub-sampled k-space patches using a deep convolutional neural network (ConvNet) to produce corresponding fully-sampled k-space patches; (f) assembling the fully-sampled k-space patches together with each other and with the sub-sampled multi-channel k-space data U to form a fully-sampled k-space data V, (g) transforming the fully-sampled k-space data V to image space using the model adjoint A.sub.adj operation to produce an image domain MRI image.
2. The method of claim 1 wherein processing the sub-sampled k-space patches using a deep convolutional neural network (ConvNet) to produce corresponding fully-sampled k-space patches comprises: processing each k-space patch u.sub.i of the sub-sampled k-space patches separately and independently from other patches to produce a corresponding fully-sampled k-space patch v.sub.i, thereby allowing for parallel processing.
3. The method of claim 2 wherein processing each k-space patch u.sub.i to produce a corresponding fully-sampled k-space patch v.sub.i comprises: applying the k-space patch u.sub.i as input to the ConvNet to infer an image space bandpass-filtered image y.sub.i, wherein the ConvNet comprises repeated de-noising blocks and data-consistency blocks; estimating the fully-sampled k-space patch v.sub.i from the image space bandpass-filtered image y.sub.i using the imaging model A and a mask matrix.
4. The method of claim 3 wherein each of the de-noising blocks comprises transforming k-space patch data to image space bandpass-filtered image data and passing the image space bandpass-filtered image data through multiple convolution layers to produce de-noised image space bandpass-filtered image data; wherein each of the data-consistency blocks comprises passing the de-noised image space bandpass-filtered image data through the imaging model A to produce known k-space patch data; wherein applying the k-space patch u.sub.i as input to a ConvNet to infer an image space bandpass-filtered image y.sub.i further comprises applying masks and a window function to k-space patch data, and passing k-space patch data through the adjoint model to produce image space bandpass-filtered image data.
5. The method of claim 4 wherein the multiple convolution layers are two-dimensional convolution layers, or three-dimensional convolution layers.
6. The method of claim 1 wherein the sub-sampled multi-channel k-space data U, sub-sampled k-space patches, fully-sampled k-space patches, fully-sampled k-space data V, and image domain MRI image are all two-dimensional data or are all three-dimensional data.
7. The method of claim 1 wherein estimating an imaging model A comprises including motion information and off-resonance de-phasing in the imaging model.
8. The method of claim 1 wherein estimating an imaging model A comprises including non-Cartesian sampling trajectories in the imaging model.
Description
BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
[0020]
[0021]
[0022]
[0023]
[0024]
DETAILED DESCRIPTION OF THE INVENTION
[0025] According to an embodiment of the invention, training and inference will all be performed on localized patches of k-space.
[0026] The reconstruction relies upon the estimation and application of imaging acquisition model A.
[0027] For each localized k-space patch, the goal of reconstruction is to solve the following inverse problem:
u.sub.i=M.sub.iA(e.sup.i2k.sup.
where u.sub.i is a selected k-space patch with its center pixel at k-space location k.sub.i, M.sub.i is a mask matrix, and y.sub.i is image-space data that is bandpass-filtered at frequency k.sub.i corresponding to the k-space patch u.sub.i. The imaging model A transforms the desired image-space data y.sub.i to the k-space (measurement) domain using sensitivity profile maps S and a Fourier transform . Sensitivity maps S are independent of the k-space patch location and can be estimated using conventional algorithms, such as ESPIRiT. Since S is set to have the same image dimensions as the k-space patch, S is faster to compute and have a smaller memory requirement in this bandpass formulation.
[0028] After the imaging model A transforms the data to the k-space domain, matrix M.sub.i is applied to mask out the missing points (due to subsampling) from the k-space patch u.sub.i. When selecting the k-space patch u.sub.i with its center pixel at k-space location k.sub.i, a phase is induced. To remove the impact of this phase when solving the inverse problem, the phase is modeled separately as e.sup.i2k.Math.x where x is the corresponding spatial location of each pixel in y.sub.i. This phase is applied through an element-wise multiplication, denoted as *.
[0029] The inverse problem of Eq. 1 can be solved to estimate the image space bandpass-filtered image data .sub.i using any standard algorithm for inverse problems with a least squares formulation with a regularization function R (y.sub.i) and regularization parameter to help constrain the problem:
In Eq. 2, we introduce a windowing function W to avoid Gibbs ringing artifacts. The model A includes sensitivity maps S that can be considered as a element-wise multiplication in the image domain or a convolution in the k-space domain. This window function also accounts for the wrapping effect of the k-space convolution when applying S in the image domain. Alternatively, the imaging acquisition model A can be applied in the k-space domain as convolutions. These k-space approaches include GRAPPA and SPIRiT. However, these approaches reconstruct y.sub.i as a multi-channel image and increases the number of channels for the regularization function R(.). In the corresponding deep neural network formulation of these approaches, the increase in number of channels will also increase the number of channels as the initial input to the neural network.
[0030] Though Eqs. 1 and 2 are set up to solve for y.sub.i which is a bandpass-filtered version of the final image, the final goal is to estimate the missing data points v.sub.i that were not originally measured. After estimating for y.sub.i, the missing points can be estimated using the modified forward imaging model as
v.sub.i=M.sub.i.sup.cA(e.sup.i2k.Math.x*y.sub.i)(3)
where M.sub.i.sup.c masks out the measured points and leaves the points that were not originally measured.
[0031] Incorporating a strong prior in the form of a regularization function has been demonstrated to enable high image quality despite high subsampling factors. In compressed sensing, the sparsity of the image in a sparsifying transform domain, such as spatial Wavelets or finite differences, can be exploited to enable undersampling factors of over 8 times Nyquist rates. Even though the problem formulation is similar to applying Wavelet transforms, directly enforcing sparsity in that domain may not be the optimal solution and the regularization parameter for each k-space location must be tuned.
[0032] To avoid the problems above with existing approaches, instead of solving Eq. 1 using a standard algorithm, the techniques of the present invention apply developments in deep convolutional neural networks (ConvNets). A key insight is that the ConvNets can be trained to rapidly solve the many small inverse problems in a feed-forward fashion. Based on the input k-space patch, the ConvNet is sufficiently flexible to adapt to solve the corresponding inverse problem, as outlined above with reference to
[0033] In our experiments, we used 6464 overlapping k-space patches. To avoid artifacts from the windowing function and from edge effects, the center 4444 of the output is inserted into the final k-space image 116. In the areas of overlap, outputs are averaged for the final solution.
[0034] The reconstruction pipeline is summarized in Algorithm 1.
[0035] A method for magnetic resonance imaging (MRI) using this reconstruction technique is shown in the flowchart of
TABLE-US-00001 Algorithm 1 Reconstruction pipeline Input: Set of k-space patches u.sub.i of full k-space image U with corresponding k-space location k.sub.i for the center pixel of each patch. U is subsampled (has missing points). Output: Reconstructed k-space image V 1: Estimate model A 2: V U {Initialize V with known measurements} 3: for all u.sub.i at k.sub.i do 4: y.sub.i G(u.sub.i, k.sub.i, A) {Inference using ConvNet G(.)} 5: v.sub.i M.sub.i.sup.cA (e.sup.i2k.Math.x * y.sub.i) {Estimate missing data points} 6: Insert v.sub.i into V 7: end for
k-space data V in step 510, and in step 512 the fully-sampled k-space data V is transformed to image space using the model adjoint A.sub.adj operation to produce an image domain MRI image.
[0036] The processing of the sub-sampled k-space patches in 508 processes each k-space patch u.sub.i of the sub-sampled k-space patches separately and independently from other patches to produce a corresponding fully-sampled k-space patch v.sub.i, thereby allowing for parallel processing. Each k-space patch u.sub.i is applied as input to the ConvNet to infer an image space bandpass-filtered image y.sub.i. The fully-sampled k-space patch v.sub.i is estimated from the image space bandpass-filtered image y.sub.i using the imaging model A and a mask matrix.
[0037] According to embodiments of the present invention, the inverse problem of Eq. 2 is solved with a convolutional neural network (ConvNet), denoted as G(.) in Algorithm 1 and
[0038] The de-noising block is composed of 2D convolution layers. The real and imaginary components of the complex data are treated as two separate channels. The input is a bandpass-filtered image of dimensions NN2. The input is passed through an initial convolution layer with 33 kernels that expands the data to 128 feature maps. The data is then passed through 5 layers of repeated 33 convolution layers with the same number of 128 feature maps. A final 33 convolution layer combines the 128 feature maps back to the 2 feature maps of real and imaginary components. Additionally, the initial input is added back to the output of the convolution layers. After each of the convolution layer except the last one, the data is passed through a batch normalization layer (BN) and a Rectified Linear Unit layer (ReLU). No normalization or activation layer is applied at the last layer to ensure that the sign (positive or negative) of the data is perserved. The input data for k-space patch u.sub.i to the k-th de-noising block R.sub.k is denoted as y.sub.i.sup.k. The output of the de-noising block is denoted as y.sub.i.sup.k+:
y.sub.i.sup.k+=R.sub.k(y.sub.i.sup.k).(4)
[0039] The data-consistency block enforces consistency with the measured data points. This block is important to ensure that the final reconstructed image agrees with the measured data points to minimize the chance of hallucination. More specifically, the data y.sub.i.sup.k+ after the k-th de-noising block is passed through the forward model to transform the data into the measurement (k-space) domain:
u.sub.i.sup.k=A(e.sup.i2k.sup.
The known measured points u.sub.i are inserted into the correct k-space locations, and then multiplied by the window function W:
u.sub.i.sup.k+1=W(M.sub.i.sup.cu.sub.i.sup.k+M.sub.iu.sub.i)(6)
The data is then passed through the adjoint model to transform the data back to the image domain:
y.sub.i.sup.k+1=e.sup.i2k.sup.
Here, A.sub.adj denotes the adjoint to A.
[0040] The two blocks, de-noising and data-consistency, are repeated. The weights in the convolution layers in the de-noising block can be kept constant for each repeated block or varied. In our experiments, we repeat the two blocks for 8 iterations and allow the weights to vary for each block to allow for more flexibility in the network.
[0041] We now turn to discussing issues of computational implementation.
[0042] To solve the inverse problem of Eq. 1, iterative algorithms are typically used. During each iteration, element-wise multiplication and addition are performed. Additionally, the inverse and forward multi-dimensional Fourier transform is performed. Despite advanced algorithmic developments, this Fourier transform is still the most computationally expensive operation. For the conventional approach of reconstructing the entire 2D image at once, each 2D Fourier transform requires O (N.sub.zN.sub.y log(N.sub.yN.sub.z)) operations for an image of dimensions N.sub.y N.sub.z.
[0043] According to the techniques of the present invention, the inverse problem is only applied for localized patches of k-space; thus, all operations including the Fourier transform are performed with smaller image dimensions. Thus, this patch-based approach significantly reduces the amount of computation. For example, given an initial image dimensions of N.sub.y=256 and N.sub.z=256, we aim to perform the reconstruction as solving the inverse problem for patches of dimensions 6464. In such a formulation, we effectively reduce the computation for the Fourier transform by over 21 fold.
[0044] In embodiments of the present invention, we can further accelerate the reconstruction procedure in two ways. First, the reconstruction of each individual k-space patch can be performed independently. This property enables the ability to parallelize the reconstruction process. Therefore, the entire reconstruction can be performed in the time in takes to reconstruct a single patch which further leverages the benefit of applying the Fourier transform operator on smaller image dimensions. Second, conventional iterative approaches to solve Eq. 1 requires an unknown number of iterations for convergence and the need to empirically tune the regularization parameter for each type of scan. In the deep learning approach of the present invention, on the other hand, the number of iterations is fixed, and the network is trained to converge to an adequate solution in the given number of iterations. Further, the need to empirically tune the regularization parameter and step sizes are eliminated as these parameters are effectively learned through the given training examples.
[0045] For purposes of illustration, we now provide examples of training and reconstruction using real data.
[0046] Volumetric abdominal images were acquired using gadolinium-contrast-enhanced MRI with a 3T scanner (GE 750 Scanner) and a 32-channel cardiac coil array. Free-breathing T1-weighted scans were collected from 301 pediatric patients using a 1-2 minute RF-spoiled gradient-recalled-echo sequence with pseudo-random Cartesian view-ordering and intrinsic navigation. For the Cartesian sampling trajectory, data were fully sampled in the k.sub.x direction (spatial frequency in x) and were subsampled in the k.sub.y and k.sub.z directions (spatial frequency in y and z). The raw imaging data was first compressed from the 32 channels to 6 virtual channels using a singular-value-decomposition-based compression scheme. The datasets were modestly subsampled with a reduction factor of 1 to 2, and the datasets were first reconstructed using parallel imaging with ESPIRiT and compressed sensing with spatial wavelets. Using the motion measured with the intrinsic navigation, respiratory motion was suppressed by weighting each data point according to the degree of motion corruption. This initial reconstruction was performed using the Berkeley Advanced Reconstruction Toolbox (BART).
[0047] For training, all volumetric data were first transformed into the hybrid (x, k.sub.y, k.sub.z)-space. Each separate x-slice was considered as a separate data sample. The dataset was divided by patient: 229 patients for training (44006 slices), 14 patients for validation (2688 slices), and 58 patients for testing (11135 slices). Thirty six different sampling masks were generated using variable density poisson disc sampling with reduction factors ranging from 2 to 9 with a fully sampled calibration region of 1010 in the center of the frequency space. During training, data was augmented by applying a randomly selected sampling mask and randomly flipping the data in y and in z. Sensitivity maps for the data acquisition model were estimated using ESPIRiT. For training, the Adam optimizer was used with .sub.1=0.9, .sub.2=0.999 and a learning rate of 0.001.
[0048] The image grid of
[0049] The final results of reconstruction using techniques of the present invention are compared with state-of-the-art compressed sensing with parallel imaging in the image grid of
[0050] The embodiments describe above are intended as concrete examples to illustrate the general principles of the invention as applied to specific implementations. Those skilled in the art will readily appreciate based on the teachings of the present application that many alternatives and variations of embodiments are possible.
[0051] The techniques of the present invention may be implemented on any standard MRI apparatus, suitable modified to reconstruct images in accordance with the techniques described here.
[0052] Different loss functions can be used for training to improve image accuracy and sharpness. These loss functions include structural similarity index metric (SSIM), l.sub.2 norm, l2 norm, and combinations of the different functions. Furthermore, the network can be trained using an adversarial network in a generative adversarial network structure.
[0053] Embodiments of the invention allows for flexibility in using different neural network structures that are used to reconstruct each frequency band. These neural network structure can include residual networks (ResNets), U-Nets, autoencoder, recurrent neural networks, and fully connected networks.
[0054] Embodiments of the invention can be modified to apply different and/or independent networks for each frequency band. For instance, one network can be trained and applied for frequency bands at lower spatial frequencies, and a different network can be trained and applied for frequency bands at higher spatial frequencies.
[0055] Additional information (such as the patch location, subsampling factor, anatomy) can be incorporated as additional inputs to the convolutional neural network.
[0056] Embodiments of the invention also allows for flexibility in modifying the imaging model used. The imaging model may include off-resonance information, signal decay model, k-space symmetry with homodyne filtering, and arbitrary sampling trajectories (radial, spiral, hybrid encoding, etc.).
[0057] Embodiments of the invention can be extended to multi-dimensional space that may include volumetric space, cardiac-motion dimension, respiratory-motion dimension, contrast-enhancement dimension, time dimension, diffusion direction, velocity, and echo dimension.
[0058] Embodiments of the invention can be used in conjunction with conventional image reconstruction methods. The results of the network can be used to initialize iterative reconstruction techniques. The results of the network can be applied for specific areas of the measurement domain: such as the center of k-space for improved data calibration for methods like parallel imaging.
[0059] Embodiments of the invention can be used to parallelize detection and correction of corrupt measurement values on a patch-by-patch basis.
[0060] The results from embodiments of the invention can also be passed through another deep neural network to further improve reconstruction accuracy.