Method for Estimating Systematic Imperfections in Medical Imaging Systems with Deep Learning
20200300957 ยท 2020-09-24
Inventors
- Feiyu Chen (Mountain View, CA)
- Christopher Michael Sandino (Menlo Park, CA)
- Joseph Yitan Cheng (Santa Clara, CA)
- John M. Pauly (Stanford, CA)
- Shreyas S. Vasanawala (Stanford, CA)
Cpc classification
G01R33/56554
PHYSICS
G01R33/5608
PHYSICS
A61B5/055
HUMAN NECESSITIES
G01R33/56518
PHYSICS
G01R33/4818
PHYSICS
G01R33/56572
PHYSICS
International classification
G01R33/565
PHYSICS
G01R33/561
PHYSICS
A61B5/055
HUMAN NECESSITIES
G01R33/58
PHYSICS
Abstract
A method for magnetic resonance imaging (MRI) includes steps of acquiring by an MRI scanner undersampled magnetic-field-gradient-encoded k-space data; performing a self-calibration of a magnetic-field-gradient-encoding point-spread function using a first neural network to estimate systematic waveform errors from the k-space data, and computing the magnetic-field-gradient-encoding point-spread function from the systematic waveform errors; reconstructing an image using a second neural network from the magnetic-field-gradient-encoding point-spread function and the k-space data.
Claims
1. A method for magnetic resonance imaging (MRI) comprising: acquiring by an MRI scanner undersampled magnetic-field-gradient-encoded k-space data; performing a self-calibration of a magnetic-field-gradient-encoding point-spread function using a first neural network to estimate systematic waveform errors from the k-space data, and computing the magnetic-field-gradient-encoding point-spread function from the systematic waveform errors; reconstructing an image using a second neural network from the magnetic-field-gradient-encoding point-spread function and the k-space data.
2. The method of claim 1 wherein the magnetic-field-gradient-encoded k-space data is wave-encoded k-space data.
3. The method of claim 1 wherein the magnetic-field-gradient-encoding point-spread function is a wave-encoding point-spread function.
4. The method of claim 1 wherein acquiring the undersampled magnetic-field-gradient-encoded k-space data comprises using a magnetic-field-gradient-encoded single shot fast spin echo sequence with variable density sampling.
5. The method of claim 1 wherein the systematic waveform errors comprise calibrated gradient time delay and isocenter location shift of magnetic-field-gradient-encoding gradients.
6. The method of claim 5 wherein the systematic waveform errors further comprise a scaling factor defining a ratio of actual and theoretical magnetic-field-gradient-encoding gradient amplitudes.
7. The method of claim 1 wherein performing a self-calibration of the magnetic-field-gradient-encoding point-spread function comprises using the first neural network to estimate systematic waveform errors from the k-space data, a theoretical maximum magnetic-field-gradient-encoding gradient amplitude, and a theoretical isocenter location.
8. The method of claim 1 wherein the second neural network used for reconstructing the image comprises multiple steps, wherein each step comprises a gradient update and a proximal step with a learned regularization network operator with trained parameters.
Description
BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
[0018]
[0019]
[0020]
[0021]
[0022]
[0023]
[0024]
[0025]
[0026]
[0027]
[0028]
[0029]
[0030]
DETAILED DESCRIPTION OF THE INVENTION
[0031] Embodiments of the present invention apply to various types of magnetic field gradient encoding. For the purposes of definiteness, the following description will focus primarily on embodiments using wave encoding.
[0032] Data Acquisition with a Wave-Encoded SSFSE Sequence
[0033] According to one embodiment of the invention, a wave-encoded single-shot fast spin echo (SSFSE) sequence with variable refocusing flip angles and variable-density sampling is used to acquire under-sampled wave-encoded k-space data. To achieve wave encoding, a sinusoidal wave-encoding gradient pulse sequence was played out on the phase encoding (PE) gradient axis during the readout of each frequency-encoding (FE, k.sub.x) line.
[0034] Variable-density sampling and variable refocusing flip angle were used in the sequence to enable full-Fourier acquisitions.sup.6. The sampling pattern contains pseudo-random sampling PE locations and a central coil calibration region of around 20 PE views.
PSF Calibration Using Data-Driven Estimation of Systematic Imperfections
[0035] Wave encoding usually requires waveform and PSF calibration due to systematic imperfections, such as gradient delays, eddy currents, and inaccurate isocenter locations. In practice, to correct this effect to the first order, a scaling factor defining the ratio between the actual and theoretical wave-encoding gradient amplitudes, a gradient delay time t, and an isocenter location shift y are introduced into the ideal waveform g.sub.y0(t). This modified waveform is then used to correct the wave-encoding PSF.sup.6. With these parameters in consideration, the actual gradient is .Math.g.sub.y0(tt), and the actual wave-encoding PSF can be expressed as
PSF[n,k]=exp(i.Math..sub.0.sup.[n].Math.g.sub.y0(tt)dt.Math.(y[k]y)),[1]
where is the gyromagnetic ratio, sampling time [n] (from the beginning of the readout) and spatial location y[k] define the two dimensions of the PSF with n defined as the index along the frequency encoding direction and k defined for phase encoding, and , t, and y are three parameters for correcting PSF errors due to systematic imperfections. Previously, systematic-imperfection-related parameters were determined by maximizing the normalized gradient of reconstructed zero-padding images with an iterative Nelder-Mead simplex method.sup.23.
[0036] According to embodiments of the invention, we replaced this iterative optimization process with a deep neural network. In our imaging systems, we observed a constant scaling factor of 0.995. Therefore, the scaling factor was kept constant in this study. In this case, the entire problem of determining systematic imperfection parameters can be described as
{circumflex over (t)},=G.sub.0(s,g.sub.max,y.sub.iso|.sub.0),[2]
where G.sub.0 is a trained neural network with three inputs: (1) the raw under-sampled, coil-combined, and wave-encoded k-space s, (2) the theoretical maximum wave-encoding gradient amplitude g.sub.max, and (3) the theoretical isocenter location y.sub.iso. Coil combination was performed by averaging the signal from all coil channels for fast computation. Both values of g.sub.max and y.sub.iso are determined based on the scan prescription. {circumflex over (t)} and represent the calibrated time delay and the isocenter shift of the wave-encoding gradients. Trained network parameters are denoted as .sub.0. There are 42,635,810 trainable parameters in .sub.0 in total.
[0037] The network architecture of G.sub.0 is illustrated in
[0038] The first step 208 converts complex numbers into real numbers by stacking the real and imaginary components together in the feature dimension. The second step 210 contains 2D operations. This step uses 2D convolutional operations and 2D pooling operations to extract k-space-domain features of the input under-sampled wave-encoded k-space. The third step 212 contains 2D to 1D operations that flatten the output of the second step 210 into a one-dimensional vector and concatenate this vector with the input maximum wave-encoding gradient amplitude g.sub.max 202 and the input theoretical isocenter location y.sub.iso 204. After two fully-connected operations 212, the network outputs {circumflex over (t)} and 206, and these two parameters are used to generate the actual wave-encoding PSF according to Eq. 1.
Image Reconstruction Using Deep Neural Networks and Gradient Updates
[0039] With the calibrated wave-encoding PSF determined as described above, a data-driven reconstruction using unrolled networks is used to recover the image. Each step of the data-driven reconstruction can be formulated as a gradient update m.sup.(k)2tA.sup.H(Am.sup.(k)s) and a proximal step with a learned regularization network operator G.sub.1.sup.(k) with trained parameters .sub.1.sup.(k). The entire step can be expressed as
m.sup.(k+1)=G.sub.1.sup.(k)(m.sup.(k)2tA.sup.H(Am.sup.(k)s)|.sub.1.sup.(k)),[3]
where m.sup.(k+1) and m.sup.(k) refer to the output and input images of the k.sup.th step. m.sup.(k+1) is also the input of the (k+1).sup.th step. 2t refers to the step size of the gradient update.sup.17. In this embodiment, four steps of iterations in Eq. 3 were used. The entire network is denoted as G.sub.1, with 1,247,044 trainable parameters. s is the partially acquired (i.e., undersampled) wave-encoded k-space data, and A describes the wave-encoding model. As described previously, the wave-encoding model can be expressed as A=D.Math..sub.y.Math.PSF.Math.
.sub.x.Math.E. In this model, D is the k-space sampling operator, E is the coil sensitivity operator, PSF is the calibrated wave-encoding PSF, and
.sub.y and
.sub.x are the Fourier transform operators in PE and FE directions. In this embodiment, PSF was computed using the data-driven approach described above.
[0040] The implementation of the sequence of G.sub.1.sup.(k) operators is shown in
Network Training and Model Deployment
[0041] To train the PSF-calibration network G.sub.0, conventional optimization-based self-calibration approach.sup.6 are used to generate ground truth labels of the gradient delay t and the isocenter shift y. Network parameters .sub.0 in G.sub.0 are minimized through 310.sup.6 steps of stochastic gradient descent with an Adam optimizer.sup.26, a learning rate of 0.005, a batch size of 20, and l.sub.1 loss.
[0042] To train the reconstruction networks in G.sub.1, conventional PICS reconstruction.sup.6 in the BART toolbox.sup.27 is used to generate ground truth labels of the image. Network parameters .sub.1 in G.sub.1 are minimized through 2.510.sup.6 steps of stochastic gradient descent with an Adam optimizer, a learning rate of 0.005, a batch size of 4, and l.sub.1 loss. Coil sensitivity maps are estimated directly using zero-padding reconstructions and SENSE.sup.14 models based on calibrated wave-encoding PSFs.
[0043] The networks G.sub.0 and G.sub.1 are trained separately on 15,783 2D wave-encoded SSFSE abdominal images on GE MR750 3T scanners collected with Institutional Review Board approval. After training of the networks is done, this data-driven pipeline can be clinically deployed and used to perform image reconstruction of wave-encoded SSFSE with constant and low computational cost.
[0044] The complete pipeline of the self-calibration and reconstruction method is illustrated in
{circumflex over (m)}=G.sub.1(s,f.sub.PSF(G.sub.0(s,g.sub.max,y.sub.iso|.sub.0))|.sub.1),[4]
where {circumflex over (m)} is the reconstructed image, f.sub.PSF is a function that generates the calibrated wave-PSF based on Eq. 1, and s is the partially acquired (i.e., undersampled) wave-encoded k-space data. As shown in the
[0045] In the PSF calibration stage 302, the conventional technique includes a theoretical wave-encoding PSF 310 and iterative self-calibration 312. According to the present invention, these are replaced with a trained neural network G.sub.0 314 to output wave-encoding gradient delays and isocenter shifts, {circumflex over (t)} and , as described above in relation to
[0046] In the reconstruction stage 306, the conventional iterative PICS reconstruction 318 is replaced with data-driven reconstruction 320, as described above in relation to
Clinical Scanning
[0047] To test the method, clinical abdominal scanning was performed on 29 consecutive adult patients (18 males, 11 females, ranging from 24 to 77 years) on a 3T MRI scanner using a 32-channel torso coil and a 2D multi-slice wave-encoded SSFSE imaging sequence. The acquisition plane was axial and PE direction was anterior/posterior. Each subject was asked to breath hold during each acquisition period. Fat-suppression was incorporated with a spectral-spatial-selective pulse. Field of view was optimized to each patient's anatomy (30-42 cm). Additional parameters are shown in Table 1. All images were reconstructed using (1) conventional PSF self-calibration and PICS reconstruction using a combination of C/C++(the BART software package.sup.27) and Python with 50 iterations and an l.sub.1-wavelet regularization coefficient of 0.002, and (2) an embodiment of the data-driven self-calibration and reconstruction method of an embodiment the invention, implemented in TensorFlow.
[0048] Representative images containing the liver are shown in
[0049] Representative images containing the kidneys are shown in
[0050] Representative images of other regions of interest are shown in
[0051] Mean difference of the gradient delays between the data-driven approach and the conventional approach was 0.0497 msec. Mean difference of the isocenter shifts between the data-driven approach and the conventional approach was 0.1006 pixel.
Individual Assessments of Image Quality
[0052] Reconstructed clinical images were independently evaluated using a semi-quantitative grading system that rated noise, contrast, sharpness, general artifacts other than motion-related artifacts, and confidence of detecting liver lesions. The scores were predetermined on a scale from 2 to 2 (Table 2). Positive values favor the data-driven method of the present invention, and negative values favor the conventional self-calibration and PICS reconstruction. Three readers independently scored each pair of the reconstructed images in a blinded, randomized order.
[0053] Wilcoxon signed-rank tests were used to test the null hypothesis that there was no significant difference between the conventional self-calibration and reconstruction approach and the data-driven approach of the invention for each reader and their average scores. Inter-observer variability was assessed utilizing a Fleiss' kappa statistic. The kappa coefficients were interpreted as almost perfect (0.8-1), substantial (0.6-0.8), moderate (0.4-0.6), fair (0.2-0.4), slight (0-0.2), and poor (<0). A two-tailed P value of under 0.05 was considered as statistical significance. The Fleiss' kappa statistic for the data-driven reconstruction of the invention vs. the conventional comparison indicated fair agreement among the readers in image sharpness (0.21) and confidence of detecting liver lesions (0.24), and slight agreement among the readers in perceived noise (0.13), general artifacts (0.09), and image contrast (0.04).
[0054] Results of individual assessments of image quality are shown in
[0055] Non-inferior image sharpness, artifacts, and diagnosis confidence for the data-driven approach of the invention were observed by all three readers. Non-inferior image quality means the approach of the invention has significantly improved image quality with P<0.05 or non-significant difference in image quality (with P values no less than 0.05) when compared to the conventional approach. Pairwise comparison of the average scores from the readers demonstrated significance in improved image sharpness, reduced artifacts, and confidence of detecting liver lesions, with mean scores of 0.23, 0.22, and 0.41, and P values of 0.001, 0.04, and 0.0003.
Evaluation of Online Computation Time
[0056] Online computation time was recorded for each patient scan using both conventional iterative self-calibration and PICS reconstruction method and the data-driven calibration and reconstruction method of an embodiment of the invention under identical hardware settings with GPU-optimized computations (two Intel Xeon CPU E5-2670 v3 @ 2.30 GHz CPUs with 24 cores each, 256 Gb RAM, and two NVIDIA TITAN X GPUs). The ratio of the average computation time between these two approaches was calculated. A t-test was performed to test the null hypothesis that there is no significant difference between the computation time of the conventional approach and the approach of the invention. A two-tailed P value of under 0.05 was considered as statistical significance.
[0057] An image reconstruction time decrease was consistently achieved in all 29 cases as shown in
[0058] Embodiments of the invention provide a data-driven calibration and reconstruction method for reconstructing clinically-relevant images of wave-encoded SSFSE imaging. By replacing iterative wave-encoding PSF calibration and PICS reconstruction with trained neural network models, the approach of the invention brought the reconstruction time closer to the acquisition duration, reducing the lags and queues in the clinical workflow. At the same time, the approach of the invention achieved reduction in perceived noise while preserving the contrast and sharpness of the current iterative approach. This approach also demonstrated a potential in reducing ghosting artifacts caused by limited field-of-views.
[0059] The invention may be implemented with different network architectures. In the self-calibration stage, the goal was to output two systematic imperfection parameters, i.e., the gradient delay and the isocenter shift between theoretical locations and actual locations. This type of output requires the neural network to reduce layer size as the layer goes deeper to extract features in the k-space domain. In the reconstruction stage, a five-layer neural network was used as a regularization function along with gradient updates. This relatively shallow design may also reduce the chance of overfitting and generating hallucinations in the early stage of the data-driven reconstruction. In the last step, a u-net was designed to improve the capacity of the network in reducing general artifacts and improving the signal-to-noise ratio. There are in total four iterations of gradient update blocks and neural network blocks. The number of iterations was empirically chosen according to a previous study.sup.28 to achieve a tradeoff between computational speed and reconstruction performance. Other embodiments may use other architectures to achieve other balances in tradeoffs.
[0060] The calibration stage and the reconstruction stage can be trained together or separately. In tests of the embodiments presented above, training of the calibration network took 148.8 hours and training of the reconstruction network took 2331.3 hours on a single NVIDIA GTX 1080Ti GPU. Training these two stages separately improves the flexibility of model deployment, as these two stages can be upgraded independently in clinical systems. Separate training may also benefit the training speed, as there are fewer unknown weights to learn in each training step, compared with joint training of all networks. At the same time, it is more convenient to check the correctness of the imperfection outputs and terminate the training of either the calibration network or the reconstruction networks when these networks are trained separately. Inaccurate PSF mainly results in ghosting artifacts near the edges of the imaging target. Therefore, improvements of PSF accuracy will reduce the level of ghosting artifacts. In these tests, ghosting artifacts due to PSF inaccuracy were hardly visible. Accuracy of the data-driven reconstruction model will affect most aspects of the final image, including sharpness, noise level, and ghosting artifacts. Therefore, the reconstruction model usually plays a more important role in the final image quality than the calibration model.
[0061] The data-driven approach of the invention reduces the total online reconstruction time by performing straightforward computations using the trained networks. In conventional calibration of the wave-encoding PSF, we notice an instability in the number of required optimization steps before convergence. In our tests, the conventional approach was been optimized for parallel computation among multiple slices. There are 48 CPU cores on our computation hardware; therefore, when the number of slices is greater than 48, the total computation time nearly doubles because a second round of computation is required for slices other than the first 48 slices. When the number of acquired slices is no larger than 48, we saw computation times of around 80 sec. When there were more than 48 slices, the computation time was around 150 sec using the conventional approach. This factor increases the instability of reconstruction time for the conventional approach. For the data-driven calibration approach, the number of computations is fixed after the models are trained, thus this approach has a relatively stable computation time. Similar behaviors can be observed in the reconstruction stage. For different number of slices (smaller or greater than 48), the computation time is relatively stable, as the reconstruction of multiple slices can be parallelized on the GPU. This time increases at a scale of 1-10 sec with increasing number of slices. The computation time of the approach of the invention in different reduction factors is also stable, as this time is only related to the number of operations in the model.
[0062] The data-driven approach of the invention reduced the perceived noise in reconstructed images by learning an optimal regularization among a large number of existing datasets. This observation is consistent with a previous study.sup.23. The improvement in perceived signal-to-noise ratio may be attributed to the diversity of the training dataset. Since our training dataset contains scans with different parameters, conditions, and thus different noise level, the trained model learns an average reconstruction of this large amount of reconstructions. Therefore, compared with conventional PICS reconstruction with a fixed regularization parameter, the resultant data-driven reconstruction achieves a more uniform perceived signal-to-noise ratio. Image-domain convolutions may also contribute to the reduction in noise level, as these operations tend to average the noise over image patches defined by the receptive field of neural networks.
[0063] The approach of the invention is able to capture small structures and maintain comparable image sharpness and contrast with respect to the conventional iterative reconstruction. This capability can be attributed to the unrolled pipeline used in the data-driven reconstruction of the invention. This pipeline contains four steps of gradient updates, which promote the data consistency between the reconstructed images and the acquired raw signal. Accurate calibration of the gradient delays and the isocenter shifts also enables good image sharpness and structural delineation comparable to the conventional reconstruction. The approach of the invention reduced phase-wrapping artifacts, as most images in the training dataset contain no phase-wrapping artifact. Pre-selection of images for training may help improve the reconstruction performance. The removal of phase wrapping artifacts using deep neural networks also means potential loss of small lesions. In this case, gradient update blocks in the reconstruction model can help ensure good data-consistency with the partially acquired k-space. The diagnostic information provided by removing phase-wrapping artifacts needs to be further evaluated.
[0064] Gradient amplitude changes and gradient delays are typically caused by eddy currents in the gradient system of the scanner. Isocenter shifts are usually caused by gradient field inhomogeneity. Therefore, for difference scanners, these imperfections may be different, and the performance of the method of the invention with extreme imperfections still needs to be evaluated. The scaling factor of maximum wave-encoding gradient was fixed in the calibration stage based on our observations that the scaling factor was stable among the scanners we have access to within a period of approximately a month. For a different constant gradient scaling factor, it is necessary to retrain the calibration network. When unstable or different scaling factors are observed on different scanners, it is also possible to include the scaling factor as an extra output parameter in the training and prediction stage while keeping the same architecture for other parts of the network. In this case, the output size of the calibration network contains three variables. The number of iterations and the regularization coefficient were also fixed in our study. This is based on previous studies on variable-density SSFSE imagine.sup.4-6. However, for applications in other regions of interest, these parameters may be different and need to be chosen empirically.
[0065] Because the approach of the invention is data-driven, a sufficient amount of data should be used. For certain applications where it may be difficult or time-consuming to acquire these datasets.sup.29-31 training data may be simulated.
[0066] In the embodiments described above, the acquisition matrix size is fixed for the data-driven approach. There are potential solutions for reconstructing scans with varying matrix sizes, which may be used if needed. For example, the input may be cropped, padded, and/or resized to the ensure uniform matrix size.
[0067] Although the tests described above were focused on wave-encoded SSFSE applications in clinical abdominal scans with fat suppression and a reduction factor of 3.5, the method of the invention may be used for other reduction factors and in other applications. For example, other applications include Cartesian encoding, Spiral encoding, Cones encoding, Propeller encoding, Radial encoding, and other non-Cartesian encoding approachs.
[0068] In conclusion, the data-driven self-calibration and reconstruction of the invention achieves an average 2.1-fold acceleration of the online computation and reduced perceived noise while maintaining non-inferior image contrast, image sharpness, artifacts, and confidence of detecting liver lesions of standard reconstruction. Together with wave-encoded SSFSE acquisition, this imaging approach provides fast and robust T2-weighted imaging as well as a fast and efficient clinical workflow.
[0069] It should be emphasized that, although the description above has focused on wave encoding as an example of a hybrid non-Cartesian sampling technique, the methods are generally applicable for reconstructing images from general non-Cartesian k-space data using data-driven deep-learning techniques according to the same techniques. The pipeline for various types of non-Cartesian sampling techniques is generally the same as that described above for wave-encoding. The estimated parameters may be adapted for different encoding methods, and the reconstruction algorithm may be adapted for different encodings. The input for other encodings may be in other domains, e.g. the k-space (spatial-frequency domain) instead of the image domain, and the network architecture can be adapted (with different operations/layers, etc.).
REFERENCES
[0070] 1. Semelka R C, Kelekis N L, Thomasson D, Brown M A, Laub G A. HASTE M R imaging: description of technique and preliminary results in the abdomen. J Magn Reson Imaging 1996; 6:698-699. [0071] 2. Feinberg D A, Hale J D, Watts J C, Kaufman L, Mark A. Halving MR imaging time by conjugation: demonstration at 3.5 kG. Radiology 1986; 161:527-531. [0072] 3. Mugler J P. Optimized three-dimensional fast-spin-echo MRI. J Magn Reson Imaging 2014; 39:745-767. [0073] 4. Loening A M, Litwiller D V, Saranathan M, Vasanawala S S. Increased Speed and Image Quality for Pelvic Single-Shot Fast Spin-Echo Imaging with Variable Refocusing Flip Angles and Full-Fourier Acquisition. Radiology 2017; 282:561-568. [0074] 5. Loening A M, Saranathan M, Ruangwattanapaisarn N, Litwiller D V, Shimakawa A, Vasanawala S S. Increased speed and image quality in single-shot fast spin echo imaging via variable refocusing flip angles. J Magn Reson Imaging 2015; 42:1747-1758. [0075] 6. Chen F, Taviani V, Tamir J I, Cheng J Y, Zhang T, Song Q, Hargreaves B A, Pauly J M, Vasanawala S S. Self-Calibrating Wave-Encoded Variable-Density Single-Shot Fast Spin Echo Imaging. J Magn Reson Imaging 2018; 47(4):954-966. [0076] 7. Hardy C J, Giaquinto R O, Piel J E, Rohling AAS K W, Marinelli L, Blezek D J, Fiveland E W, Darrow R D, Foo T K. 128-channel body MRI with a flexible high-density receiver-coil array. J Magn Reson Imaging 2008; 28(5):1219-1225. [0077] 8. Busse R F, Brau A C, Vu A, Michelich C R, Bayram E, Kijowski R, Reeder S B, Rowley H A. Effects of refocusing flip angle modulation and view ordering in 3D fast spin echo. Magn Reson Med 2008; 60:640-649. [0078] 9. Bilgic B, Gagoski B A, Cauley S F, Fan A P, Polimeni J R, Grant P E, Wald L L, Setsompop K. Wave-CAIPI for highly accelerated 3d imaging. Magn Reson Med 2015; 73:2152-2162. [0079] 10. Gagoski B A, Bilgic B, Eichner C, Bhat H, Grant P E, Wald L L, Setsompop K.RARE/turbo spin echo imaging with Simultaneous Multislice Wave-CAIPI. Magn Reson Med 2015; 73:929-938. [0080] 11. Chen F, Zhang T, Cheng J Y, Shi X, Pauly J M, Vasanawala S S. Auto-calibrating motion-corrected wave-encoding for highly accelerated free-breathing abdominal MRI. Magn Reson Med 2017; 78(5):1757-1766. [0081] 12. Chen F, Zhang T, Cheng J Y, Taviani V, Hargreaves B, Pauly J, Vasanawala S, inventors; General Electric Co, Leland Stanford Junior University, assignee. System and method for performing wave-encoded magnetic resonance imaging of an object. U.S. patent application Ser. No. 15/481,893. 2018 May 24. [0082] 13. Cauley S F, Setsompop K, Bilgic B, Bhat H, Gagoski B, Wald L L. Autocalibrated wave-CAIPI reconstruction; Joint optimization of k-space trajectory and parallel imaging reconstruction. Magn Reson Med 2017; 78(3):1093-1099. [0083] 14. Uecker M, Lai P, Murphy M J, Virtue P, Elad M, Pauly J M, Vasanawala S S, Lustig M. ESPIRiT: an eigenvalue approach to autocalibrating parallel MRI: where SENSE meets GRAPPA. Magn Reson Med 2014; 71:990-1001. [0084] 15. Lustig M, Donoho D, Pauly J M. Sparse MRI: The application of compressed sensing for rapid MR imaging. Magn Reson Med 2007; 58:1182-1195. [0085] 16. Shi X, Ma X, Wu W, Huang F, Yuan C, Guo H. Parallel imaging and compressed sensing combined framework for accelerating high-resolution diffusion tensor imaging using inter-image correlation. Magn Reson Med 2015; 73(5):1775-1785. [0086] 17. Cheng J Y, Chen F, Sandino C, Mardani M, Pauly J M, Vasanawala S S. Compressed Sensing: From Research to Clinical Practice with Data-Driven Learning. arXiv preprint arXiv:1903.07824. [0087] 18. Cheng J Y, Chen F, Alley M T, Pauly J M, and Vasanawala S S. Highly Scalable Image Reconstruction using Deep Neural Networks with Bandpass Filtering. arXiv preprint arXiv:1805.03300. [0088] 19. Yang Y, Sun J, Li H, and Xu Z. ADMM-Net: A Deep Learning Approach for Compressive Sensing MRI. NIPS 2017; 10-18. [0089] 20. Adler J, ktem O. Learned primal-dual reconstruction. IEEE transactions on medical imaging. 2018; 37(6):1322-32. [0090] 21. Hammernik K, Klatzer T, Kobler E, Recht M P, Sodickson D K, Pock T, Knoll F. Learning a variational network for reconstruction of accelerated MRI data. Magn Reson Med 2018; 79(6):3055-3071. [0091] 22. Chen F, Taviani V, Malkiel I, Cheng J Y, Tamir J I, Shaikh J, Chang S T, Hardy C J, Pauly J M, Vasanawala S S. Variable-density single-shot fast spin-echo MRI with deep learning reconstruction by using variational networks. Radiology. 2018; 289(2):366-373. [0092] 23. Olsson D M, Nelson L S. The nelder-mead simplex procedure for function minimization. Technometrics 1975; 17:45-51. [0093] 24. Krizhevsky A, Sutskever I, Hinton G E. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems 2012; 1097-1105. [0094] 25. Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical image computing and computer-assisted intervention 2015; 234-241. [0095] 26. Kingma D P, Ba J. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980. [0096] 27. Uecker M, Ong F, Tamir J I, Bahri D, Virtue P, Cheng J Y, Zhang T, Lustig M. Berkeley advanced reconstruction toolbox. Proc. Intl. Soc. Mag. Reson. Med 2015; 23:2486. [0097] 28. Mardani M, Sun Q, Donoho D, Papyan V, Monajemi H, Vasanawala S, Pauly J. Neural proximal gradient descent for compressive imaging. In Advances in Neural Information Processing Systems 2018; 9573-9583. [0098] 29. Chen F, Cheng J Y, Pauly J M, Vasanawala S S. Semi-Supervised Learning for Reconstructing Under-Sampled MR Scans. Proc. Intl Soc Mag Reson Med 2019; 27:4649. [0099] 30. Chen F, Shi X, Chen S, Johnson E M, Chen B, Ren G, Wei X, Wang S, Ying K. Accelerated model-based proton resonance frequency shift temperature mapping using echo-based GRAPPA reconstruction. Magnetic resonance imaging 2015; 33 (2):240-245. [0100] 31. Hu Y, Levine E G, Tian Q, Moran C J, Wang X, Taviani V, Vasanawala S S, McNab J A, Daniel B A, Hargreaves B L. Motion-robust reconstruction of multishot diffusion-weighted images without phase estimation through locally low-rank regularization. Magn Reson Med 2019; 81(2):1181-1190.
TABLE-US-00001 TABLE 1 Scan Parameter Summary Refocusing flip angles 130/90/100/45 (initial/minimum/central/last) Slice thickness 5 mm Slice gap 0 Matrix 320 180 FOV 300-420 mm 240-336 mm Number of slices 35-64 Echo time (TE) 90.9-93.1 msec Repetition time (TR) 436.5-560.6 msec Total scan time 16.9-32.8 sec Bandwidth 488.2 Hz/pixel Acceleration factor for 3.5 variable density sampling Sampling pattern Variable-density sampling (51 acquired PE views) with an auto-calibration region (20 PE views) Proportion of k-space coverage 1 (full)
TABLE-US-00002 TABLE 2 Scoring Criteria for Conventional Self-Calibration and Reconstruction vs. Data-Driven Self-Calibration and Reconstruction of the Invention Favors A Favors B Score 2 1 0 1 2 Noise A with A with Equivalent B with B with decreased decreased decreased decreased graininess with graininess graininess graininess with improved without without improved diagnostic diagnostic diagnostic diagnostic capability impact impact capability Contrast A with good A with good Equivalent B with good B with good contrast contrast contrast contrast between between between between liver/spleen liver/spleen OR liver/spleen OR liver/spleen AND renal renal renal AND renal cortex/medulla cortex/medulla cortex/medulla cortex/medulla not seen in B not seen in B not seen in A not seen in A Sharpness A with A with Equivalent B with B with increased increased increased increased sharpness with sharpness sharpness sharpness with improved without without improved diagnostic diagnostic diagnostic diagnostic capability impact impact capability Artifacts A with A with Equivalent B with B with (besides decreased decreased decreased decreased cardiac motion artifacts to the artifacts artifacts artifacts to the related point of without without point of nonuniformity) improved diagnostic diagnostic improved diagnostic impact impact diagnostic capability capability Confidence of A with A with Equivalent B with B with detecting liver improved improved improved improved lesions confidence of confidence of confidence of confidence of detecting liver detecting liver detecting liver detecting liver lesions in more lesions in fewer lesions in fewer lesions in more than half of the than half of the than half of the than half of the images images images images