Methods and systems for dynamic coronary roadmapping
11707242 · 2023-07-25
Assignee
Inventors
- Theo van Walsum (Houten, NL)
- Hua Ma (Rotterdam, NL)
- Jean-Paul Aben (Limbricht, NL)
- Dennis Koehn (Voerendaal, NL)
Cpc classification
G16H50/20
PHYSICS
A61B5/349
HUMAN NECESSITIES
A61B6/5288
HUMAN NECESSITIES
A61B6/463
HUMAN NECESSITIES
A61B6/12
HUMAN NECESSITIES
A61B6/5235
HUMAN NECESSITIES
A61B6/5247
HUMAN NECESSITIES
A61B6/504
HUMAN NECESSITIES
A61B6/5205
HUMAN NECESSITIES
A61B8/483
HUMAN NECESSITIES
International classification
A61B6/00
HUMAN NECESSITIES
A61B5/00
HUMAN NECESSITIES
A61B5/349
HUMAN NECESSITIES
Abstract
Methods are provided for dynamically visualizing information in image data of an object of interest of a patient, which include an offline phase and an online phase. In the offline phase, first image data of the object of interest acquired with a contrast agent is obtained with an interventional device is present in the first image data. The first image data is used to generate a plurality of roadmaps of the object of interest. A plurality of reference locations of the device in the first image data is determined, wherein the plurality of reference locations correspond to the plurality of roadmaps. In the online phase, live image data of the object of interest acquired without a contrast agent is obtained with the device present in the live image data, and a roadmap is selected from the plurality of roadmaps. A location of the device in the live image data is determined. The reference location of the device corresponding to the selected roadmap and the location of the device in the live image data is used to transform the selected roadmap to generate a dynamic roadmap of the object of interest. A visual representation of the dynamic roadmap is overlaid on the live image data for display. In embodiments, the first image data of the offline phase covers different of phases of the cardiac cycle of the patient, and the plurality of roadmaps generated in the offline phase covers the different phases of the patient's cardiac cycle. Related systems and program storage devices are also described and claimed.
Claims
1. A method for generating an image of an object of interest of a patient, the object of interest comprising a heart, a part of the coronary tree, blood vessels or other part of the vasculature of the patient, the method comprising: i) obtaining first image data of the object of interest, wherein the first image data is acquired using an X-ray imaging modality with a contrast agent and an interventional device is present in the first image data, the interventional device being used in a procedure to treat the object of interest, wherein the first image data covers at least one cardiac cycle of the patient; ii) using the first image data to generate a plurality of roadmaps of the object of interest; iii) determining a plurality of reference locations of a tip of the interventional device in the first image data, wherein the plurality of reference locations correspond to the plurality of roadmaps of the object of interest; iv) obtaining second image data of the object of interest, wherein the second image data is acquired using an X-ray imaging modality without a contrast agent and the interventional device is present in the second image data; v) selecting a roadmap from the plurality of roadmaps; vi) determining a location of the tip of the interventional device in the second image data; vii) using the reference location of the tip of the interventional device corresponding to the roadmap selected in v) and the location of the tip of the interventional device determined in vi) to transform the roadmap selected in v) to generate a dynamic roadmap of the object of interest; and viii) overlaying a visual representation of the dynamic roadmap of the object of interest as generated in vii) on the second image data for display; wherein, in vi), the location of the tip of the interventional device in the second image data is determined by inputting the second image data to a trained machine learning network that outputs a posterior probability distribution that estimates likelihood of the location of the tip of the interventional device in the second image data given the second image data as input and using a Bayesian filtering method that equates location of the tip of the interventional device to a weighted arithmetic mean of a plurality of positions and their associated weights derived from the posterior probability distribution output by the trained machine learning network.
2. A method according to claim 1, wherein: the plurality of roadmaps of the object of interest as generated in ii) covers different phases of the cardiac cycle of the patient.
3. A method according to claim 2, wherein: the phases of the cardiac cycle of the patient are offset in time relative to a predefined reference part of the cardiac cycle of the patient.
4. A method according to claim 2, further comprising: acquiring an ECG signal while acquiring the second image data, and processing the ECG signal to determine a phase of the cardiac cycle of the patient that corresponds to the second image data; and selecting the roadmap in v) by matching the phase of the cardiac cycle of the patient for the second image data to the phase of the cardiac cycle of the patient for the selected roadmap.
5. A method according to claim 4, further comprising: processing the first image data to determine a phase of the cardiac cycle of the patient for an image frame and associating the phase of the cardiac cycle to a roadmap corresponding to the image frame.
6. A method according to claim 1, wherein: the plurality of roadmaps of the object of interest comprise a plurality of three-dimensional roadmaps.
7. A method according to claim 6, wherein: the plurality of three-dimensional roadmaps are derived from a three-dimensional model of the object of interest.
8. A method according to claim 7, wherein: the three-dimensional model of the object of interest is extracted from at least one image modality selected from the group consisting of computed tomography (CT), X-ray rotational angiography, 3D Ultrasound, or magnetic resonance imaging (MRI).
9. A method according to claim 6, wherein: the plurality of three-dimensional roadmaps are derived from two X-ray angiographic image sequences of the object of interest acquired with a contrast agent.
10. A method according to claim 6, wherein: the plurality of three-dimensional roadmaps are derived from a three-dimensional model of the object of interest and at least one X-ray angiographic image sequence of the object of interest acquired with a contrast agent.
11. A method according to claim 10, wherein: the three-dimensional model of the object of interest is extracted from at least one image modality selected from the group consisting of computed tomography (CT), X-ray rotational angiography, 3D Ultrasound, or magnetic resonance imaging (MRI).
12. A method according to claim 1, wherein: the plurality of roadmaps include information that characterizes properties of the object of interest.
13. A method according to claim 1, wherein: the plurality of roadmaps include at least one measurement for the object of interest selected from the group consisting of location and extent of vessel obstruction, diameter and area, pressure, blood velocity, fractional flow reserve, wall shear stress, vessel curvature, amount of foreshortening, location and extent and type of coronary plaque, location and extent of coronary total occlusion, or location and extent of coronary obstruction.
14. A method according to claim 1, wherein: the Bayesian filtering method involves resampling points around a position with a high weight value.
15. A method according to claim 1, wherein: the selection of the roadmap in v) is configured to compensate for cardiac motion.
16. A method according to claim 1, wherein: the operations of vii) apply a transformation to the roadmap selected in v) in order to compensate for motion between the first image data and the second image data.
17. A method according to claim 16, wherein: the motion includes breathing motion and/or cardiac motion and/or patient motion and/or table motion.
18. A method according to claim 16, wherein: the transformation comprises a rigid transformation or a non-rigid transformation to the roadmap selected in v) based on a displacement obtained from the reference location of the device corresponding to the roadmap selected in v) and the location of the device determined in vi).
19. A method according to claim 1, wherein: the visual representation of the dynamic roadmap is generating by projecting the overlay of the dynamic roadmap onto the second image data using a transparent mode, and/or dilating the dynamic roadmap and projecting the boundaries of the resultant dynamic roadmap onto the second image data; whereby the visual representation of the dynamic roadmap is configured to not obscure any instrument used to treat the object of interest.
20. A method according to claim 1, wherein: the plurality of reference locations is stored as part of the plurality of roadmaps of the object of interest.
21. A method according to claim 1, wherein: the roadmap selected in v) comprises a three-dimensional roadmap that is transformed to generate at least one dynamic roadmap for overlay on the second image data.
22. A method according to claim 21, wherein: the three-dimensional roadmap is transformed according to the viewpoint used to acquire the second image data.
23. A method according to claim 21, wherein: the second image data is acquired from a viewpoint different from the first image data.
24. A method according to claim 1, wherein: the roadmap selected in v) comprises a two-dimensional roadmap that is transformed to generate at least one dynamic roadmap for overlay on the second image data; and the first image data and the second image data are acquired from a common viewpoint.
25. A method according to claim 1, wherein: the first image data is derived by subtraction of a baseline image; and the second image data is derived by subtraction of the baseline image.
26. A method according to claim 1, wherein: the operations of iv) to viii) are repeated for successive frames of a live image sequence acquired without a contrast agent.
27. A method according to claim 1, wherein: the interventional device is selected from the group consisting of a guiding catheter, a guide wire, or other intraluminal device or instrument.
28. A method according to claim 1, further comprising: displaying the overlay of the visual representation of the dynamic roadmap of the object of interest on the second image data.
29. A method according to claim 1, wherein: the posterior probability distribution is represented in the image pixel space of the second image data.
30. A method according to claim 1, further comprising: training the machine learning network using a probability distribution derived from a known location of a tip of an interventional device in a training image dataset.
31. A method according to claim 30, wherein: the training of the machine learning network uses a catheter segmentation heatmap in combination with the probability distribution derived from a known location of a tip of an interventional device in a training image dataset.
32. A method according to claim 1, wherein: the Bayesian filtering method further involves determining random samples and associated weights that approximate the posterior probability distribution output by the trained machine learning network by applying a motion model to the random samples, wherein the motion model is estimated from adjacent image frames.
33. A method according to claim 32, wherein: the motion model is estimated using an optical flow method.
34. A method according to claim 32, wherein: the Bayesian filtering method further involves resampling the random samples and updating the associated weights in a manner that maximizes the number of effective random samples that have an actual contribution in approximating the posterior probability distribution output by the trained machine learning network.
35. A method according to claim 1, wherein: the machine learning network comprises a convolutional neural network.
36. A system for generating an image of an object of interest of a patient, the object of interest comprising a heart, a part of the coronary tree, blood vessels or other part of the vasculature of the patient, the system comprising: at least one processor that, when executing program instructions stored in memory, is configured to i) obtain first image data of the object of interest, wherein the first image data is acquired using an X-ray imaging modality with a contrast agent and an interventional device is present in the first image data, the interventional device being used in a procedure to treat the object of interest, wherein the first image data covers at least one cardiac cycle of the patient; ii) use the first image data to generate a plurality of roadmaps of the object of interest; iii) determine a plurality of reference locations of a tip of the interventional device in the first image data, wherein the plurality of reference locations correspond to the plurality of roadmaps of the object of interest; iv) obtain second image data of the object of interest, wherein the second image data is acquired using an X-ray imaging modality without a contrast agent and the interventional device is present in the second image data; v) select a roadmap from the plurality of roadmaps; vi) determine a location of the tip of the interventional device in the second image data; vii) use the reference location of the tip of the interventional device corresponding to the roadmap selected in v) and the location of the tip of the interventional device determined in vi) to transform the roadmap selected in v) to generate a dynamic roadmap of the object of interest; and viii) overlay a visual representation of the dynamic roadmap of the object of interest as generated in vii) on the second image data for display; wherein, in vi), the location of the tip of the interventional device in the second image data is determined by inputting the second image data to a trained machine learning network that outputs a posterior probability distribution that estimates likelihood of the location of the tip of the interventional device in the second image data given the second image data as input and using a Bayesian filtering method that equates location of the tip of the interventional device to a weighted arithmetic mean of a plurality of positions and their associated weights derived from the posterior probability distribution output by the trained machine learning network.
37. A system according to claim 36, further comprising: an imaging acquisition subsystem configured to acquire the first image data and the second image data, wherein the imaging acquisition subsystem uses an X-ray imaging modality.
38. A system according to claim 37, further comprising: a display subsystem configured to display the overlay of the visual representation of the dynamic roadmap of the object of interest as generated in viii) on the second image data.
39. A non-transitory program storage device tangibly embodying a program of instructions that are executable on a machine to perform operations for generating an image of an object of interest of a patient, the object of interest comprising a heart, a part of the coronary tree, blood vessels or other part of the vasculature of the patient, the operations comprising: i) obtaining first image data of the object of interest, wherein the first image data is acquired using an X-ray imaging modality with a contrast agent and an interventional device is present in the first image data, the interventional device being used in a procedure to treat the object of interest, wherein the first image data covers at least one cardiac cycle of the patient ii) using the first image data to generate a plurality of roadmaps of the object of interest iii) determining a plurality of reference locations of a tip of the interventional device in the first image data, wherein the plurality of reference locations correspond to the plurality of roadmaps of the object of interest iv) obtaining second image data of the object of interest, wherein the second image data is acquired using an X-ray imaging modality without a contrast agent and the interventional device is present in the second image data; v) selecting a roadmap from the plurality of roadmaps; vi) determining a location of the tip of the interventional device in the second image data; vii) using the reference location of the tip of the interventional device corresponding to the roadmap selected in v) and the location of the tip of the interventional device determined in vi) to transform the roadmap selected in v) to generate a dynamic roadmap of the object of interest and viii) overlaying a visual representation of the dynamic roadmap of the object of interest as generated in vii) on the second image data for display; wherein, in vi), the location of the tip of the interventional device in the second image data is determined by inputting the second image data to a trained machine learning network that outputs a posterior probability distribution that estimates likelihood of the location of the tip of the interventional device in the second image data given the second image data as input and using a Bayesian filtering method that equates location of the tip of the interventional device to a weighted arithmetic mean of a plurality of positions and their associated weights derived from the posterior probability distribution output by the trained machine learning network.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
(21)
(22)
(23)
(24)
(25)
(26)
(27)
(28)
(29)
(30)
(31)
(32)
(33)
(34)
(35)
(36)
(37)
(38)
(39)
(40)
(41)
DETAILED DESCRIPTION OF ILLUSTRATIVE EMBODIMENT(S)
(42) This present application describes method(s) and system(s) to provide a real time dynamic overlay or dynamic coronary roadmap which can be superimposed on the live X-ray fluoroscopic or angiographic image stream/sequence and thereby providing support for the clinician in improved patient treatment. The approach compensates changes in vessel shapes and cardiac motion by selecting a roadmap of the same cardiac phase through temporal alignment, and corrects the respiratory induced motion and patient motion via tracking a device.
(43) In the description of the methodology and systems below, X-ray angiography will be used to refer to an X-ray acquisition of an object of interest (e.g., a part of the vasculature or a coronary vessel tree) after administration of a contrast agent resulting in visualization of the object of interest and other objects which are radiopaque. X-ray fluoroscopy refers to X-ray image acquisition without the use of a contrast agent that and therefore contains no enhance visualization of the object of interest. The term image or image frame refers to a single image, and the term image sequence or image stream refers to a multiple images acquired over time. The X-ray angiography image sequence may contain the contrast agent administration, and image frame which precedes the contrast agent administration will contain no contrast agent and thus no enhancement of the vasculature. The X-ray angiography image sequence may comprise multiple frames covering one or more phases of the cardiac cycle. Throughout this patent application, a cardiac cycle is specific to a patent and is defined as the period in which covers one heartbeat of the patient. The cardiac cycle can be defined as the period of time between successive R-tops within the ECG signal of the patient. A phase refers to a moment (or period) of time within the cardiac cycle of the patient. Phase can be measured as an offset from R-top within the ECG signal of the patient as shown in
(44) Within this application a method is proposed to generate and display the dynamic overlay or dynamic roadmap as an overlay on the live X-ray image stream, which consists of the following elements: Construct a model of the organ of interest (such as a coronary tree) from angiographic image sequence of the organ of interest; Select a part (or frame) of the model that is aligned with the cardiac phase of the patient using electrocardiogram (ECG) measurements of the patient; Update or transform the selected model part to compensate for respiratory motion using guiding catheter tracking in live X-ray fluoroscopic or angiographic image stream of the patient; Render the resulting model part and integrate the rendered view of the model as an overlay on the live X-ray fluoroscopic or angiographic image stream of the patient.
The described dynamic coronary roadmap method runs in real-time with a graphics processing unit (GPU), and thus can be used during PCI in real clinical settings.
(45)
(46)
(47) The operations of
(48) In this example it is assumed that the imaging system has acquired and stored at least one two-dimensional image sequence of an object of interest. Any image device capable of providing two-dimensional angiographic image sequences can be used for this purpose. For example a biplane or single plane angiographic system can be used such as those manufactured, for example, by Siemens (Artis zee Biplane) or Philips (Allura Xper FD).
(49)
(50) An embodiment is now disclosed with reference to
(51) Before explanation of each step of
(52) Offline Phase:
(53) The offline phase, represented by block 100 in
(54) A device location (e.g. the catheter tip location) within the X-ray angiography sequence is obtained and associated with the roadmaps to serve as a reference point for roadmap transformation (as described in more detail by step 107 of
(55) Online Phase:
(56) During the online phase, represented by block 110 in
(57) An illustration of the dynamic roadmap process is presented in FIG. 3. Within
(58) In the following sections, the steps describes by
(59) Step 101: Retrieve Angiographic Image Data
(60) In the first step, step 101, the angiographic image data is retrieved. Within a preferred embodiment the angiographic image data represents the acquisition of an object of interest by means of X-ray angiography, resulting in an X-ray angiographic image sequence. For example a single plane or bi-plane angiographic system can be used such as those manufactured, for example, by Siemens (Artis zee Biplane) or Philips (Allura Xper FD). The object of interest is for example the heart, a part of the coronary tree, or blood vessels and by means of a contrast medium the object of interest is visualized during the acquisition. The X-ray angiography image sequence is acquired in such a way that the object of interest is clearly visible. Therefore, the C-arm can be rotated and angulated by the clinician to obtain the best projection. Preferably, the electrocardiogram (ECG) is part of the angiographic image data and is simultaneously recorded during the X-ray angiography acquisition. The ECG enables cardiac phase matching between different X-ray angiography frames or X-ray angiography acquisitions. With cardiac phase matching the object of interest displacement due to cardiac motion between image frames in the same cardiac phase is minimized.
(61)
(62) Step 102: Obtain Device Location within Angiographic Image Data
(63) There is a number of factors that can cause displacement of the object of interest or the roadmap between successive X-ray image frames within an X-ray angiographic image sequence or between different X-ray angiographic image sequences. These factors are cardiac motion due to contraction of the heart, breathing motion due to respiration of the patient and patient motion due to patient movement. Cardiac motion is compensated by matching the cardiac phase using for example the retrieved ECG signal. Breathing motion, including possible patient motion, can be compensated by identifying a reference point in the image. The reference point (represented by a visible device in the image sequence) will be used for transformation of the object of interest or the roadmap during the online phase. The reference point could be for example be the catheter tip, pacemaker or anatomical landmarks or any other object in which its motion can be correlated to the breathing motion and possible patient motion. Such a reference point can be obtained in every frame of the X-ray angiographic image sequence. Identification of the reference point in a frame might be done manually. Alternatively to a manual identification of the reference point, image processing techniques might be applied to detect objects or landmarks in medical image data for example as described by step 503 of
(64) Within a preferred embodiment, the steps to obtain the device location within an X-ray angiographic image sequence are illustrated by the flowchart of
(65)
(66) where Median.sub.(frame) represent the median of an image frame, I.sub.(frame)(n) represents a pixel n within an image frame, c represent a constant, and ω.sub.(frame) represents a weighting function of an image frame in which the likelihood of tubular structures are represented; zero means low likelihood of tubular structures and 1.0 represent a high likelihood of tubular structures. The likelihood can be calculated by for instance by applying a Frangi vesselness filter (Frangi et al., “Multiscale Vessel Enhancement Filtering”, Medical Image Computing and Computer-Assisted Intervention—MICCAI 1998 Lecture Notes in Computer Science 1496/1998:130).
(67) Another approach to detect the frame within the X-ray angiographic image sequence in which the contrast liquid enters the coronary tree is the method as taught by Ma, et al., “Fast prospective detection of contrast inflow in x-ray angiograms with convolutional neural network and recurrent neural network”, International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer (2017) 453-461. In this work, Ma et al. describes two different approaches for detection of the frame within an image sequence in which the contrast first appears. The first approach trains a convolutional neural network (CNN) to distinguish whether a frame has contrast agent or not. The second approach extracts first contrast features from images with enhanced vessel structures and the contrast frames are then detected based on changes in the feature curve using long short-term memory which is established by a recurrent neural network architecture.
(68) Next, in step 502 the cardiac cycle information is extracted from the X-ray angiographic image sequence. In case the ECG signal is part of the angiographic image data (101 of
(69) One approach to extract the cardiac cycle information is illustrated by 702 of
(70) Within
E=X{tilde over (E)}∇.sup.−1 (equation 2)
(71) where E is the D×N matrix of eigenvectors of XX.sup.T, E is the N×N matrix of eigenvectors, and ∇ is the N×N diagonal matrix of eigenvalues of X.sup.TX.
(72) Next, the post processed sequence (703) is projected on the first principal component e.sub.1 by computing
p=X.sup.Te.sub.1 (equation 3)
(73) where e.sub.1 is the first column of E representing the direction of the largest variance and p is a N×1 projection vector. So each frame in such sequence is represented by an element in vector p. The assumption underlying the described approach is that cardiac motion is the major source of variation in these sequences (703) where respiratory motion and patient motion are eliminated. Therefore, p is used as the cardiac motion surrogate representing the cardiac motion within the image sequence 701.
(74) Referring back to
(75) Another method to detect the device (e.g. the catheter tip) within frames N.sub.d is by manually identifying the catheter tip in one frame, and propagate this location to the remaining frames within N.sub.d by means of image processing techniques for example as taught Zang et al., “Cascade Attention Machine for Occluded Landmark Detection in 2D X-Ray Angiography”, Jan. 2019 IEEE Winter Conference on Applications of Computer Vision (WACV).
(76) Another method to detect the device (e.g. the catheter tip) may be performed by model based detection, or use of convolution neural networks. Detection of the catheter tip might be improved by incorporating temporal information in the preprocessing or post processing. A fully automatic catheter segmentation technique based on convolutional neural network and including temporal information is taught by Ambrosini et al., “Fully automatic and real-time catheter segmentation in X-ray fluoroscopy”, International Conference on Medical Image Computing and Computer-Assisted Intervention 2017, Springer. pp. 577-585.
(77) Another method to detect the device (e.g. the catheter tip) is by first selecting an amount of successive frames representing once cardiac cycle before the frame in which the contrast enters the coronary artery (as described by step 501 of
(78) Referring back to
(79) Step 103: Create Roadmaps
(80) Within step 103, the X-ray angiographic image sequence is processed to create roadmaps of coronary arteries for multiple phases of the cardiac cycle after the frame in which the contrast liquids enters the coronary artery. Typically, the roadmaps are created for an amount of frames covering at least one cardiac cycle.
(81) For example, the coronary artery roadmap can include a vessel model. The vessel model can represent the vessels centerlines, or the vessels boundary (contours), or a mask image which represents the vessels. Furthermore, the vessel model can also include clinically relevant information, such as for example location and percentage of vessel obstruction, and/or diameter, area, length of the vessel or the curvature of the vessel, and/or the location and amount of calcified plaque.
(82) In embodiments, the coronary artery roadmap can be created for all frames within one cardiac cycle after contrast injection. For example, a full cardiac cycle can be selected within the X-ray angiography image sequence (as a result from step 101) after the frame in which the contrast liquids enters the coronary artery (as a result of step 501 of
(83)
(84) Another approach to create the roadmap is by applying image processing skeleton techniques as for instance taught by Li et al., “Skeletonization of gray-scale image from incomplete boundaries”, Proceedings of the International Conference on Image Processing, ICIP 2008, Oct. 12-15). Another approach to create the roadmap is by utilizing deep learning techniques as for instance taught by Nasr-Esfahani et al., “Vessel extraction in X-ray angiograms using deep learning”, Conf Proc IEEE Eng Med Biol Soc. 2016 August; 2016:643-646 or as taught by Wang et al., “Vessel extraction in coronary X-ray Angiography”, Conf Proc IEEE Eng Med Biol Soc. 2005; 2: 1584-1587.
(85) Along with the N.sub.after-contrast ECG signals or the extracted cardiac cycle as a result from step 501 of
(86) Optionally, based on the generated roadmaps, quantitative image analysis can be performed to extract clinical relevant information, such as for example location and percentage of vessel obstruction, diameter and area, length of the vessel or the curvature of the vessel as for instance taught by Girasis et al., “Advances in two-dimensional quantitative coronary angiographic assessment of bifurcation lesions: improved small lumen diameter detection and automatic reference vessel diameter derivation”, EuroIntervention 2012 March; 7(11):1326-35 or as taught by Wang et al., “Vessel extraction in coronary X-ray Angiography”, Conf Proc IEEE Eng Med Biol Soc. 2005; 2: 1584-1587.
(87) Optionally, the location and amount of calcified plaque can be extracted from the X-ray angiographic image sequence as for instance disclosed in detail by the flowchart description of
(88) Finally, in step 103 the obtained device location as a result of step 102 is integrated into the created roadmaps.
(89) Step 104: Retrieve Fluoroscopic Image Data
(90) The first step in the online phase (110), is the retrieval of the X-ray fluoroscopic image data represented by step 104 of
(91) Step 105: Select Roadmap
(92) After retrieving fluoroscopic images including ECG signal as described by step 104, a roadmap from the selection of roadmaps (which are created in step 103) is selected for every single fluoroscopic image.
(93) Roadmap selection represented by step 104 in
(94) To select roadmaps and images based on ECG, a temporal mapping between X-ray images and ECG signal points needs to be built first. It is assumed that ECG signals and X-ray images are well synchronized during acquisition.
(95) In the offline phase (block 100 of
(96) In addition to this, the method of step 501 can be applied to discard images prior to contrast liquid arrival in the vessel. This will speed up the process of mapping ECG data and the image sequence.
(97) In the online phase (block 110 of
(98) To compare the ECG signals associated with the offline angiographic sequence and the online fluoroscopic image, a temporal registration of the two signals using for example cross-correlation is applied, such is taught by Kim et al., “Registration of angiographic image on real-time fluoroscopic image for image-guided percutaneous coronary intervention”, International journal of computer assisted radiology and surgery 2018:13, 203-213. The two ECG signals are first cross-correlated for every possible position on the signals, resulting in a one dimensional (1D) vector of correlation scores. The candidate frame for dynamic overlay and roadmap is then selected as the one associated with the point on the ECG of the offline angiographic sequence that is corresponding to the highest correlation score.
(99) Step 106: Track Device
(100) In the online phase (block 110 of
(101) Next a method for tracking a device is presented and as an example the tracked object is a catheter tip.
(102) Exemplary Catheter Tip Tracking
(103) The overall catheter tip tracking uses a deep learning based Bayesian filtering method and is summarized in
(104) Bayesian Filtering
(105) Bayesian filtering is a state-space approach aiming at estimating the true state of a system that changes over time from a sequence of noisy measurement made on the system as for instance described by Arulampalam et al., “A tutorial on particle filters for online nonlinear/non-Gaussian Bayesian tracking”, IEEE Transactions on signal processing 2002:50, 174-188.
(106) Bayesian filtering typically includes the following components: hidden system states, a state transition model, observations and an observation model. Let x.sub.k∈.sup.2(k={0, 1, 2, . . . }) denote the state, the location of guiding catheter tip in the k-th frame, a 2D vector representing the coordinates in the X-ray image space. The transition of the system from one state to the next state is given by the state transition model x.sub.k=f.sub.k(x.sub.k-1, v.sub.k-1), where v.sub.k-1∈
.sup.2 is an independent and identically distributed (i.i.d.) process noise, f.sub.k:
.sup.2×
.sup.2.fwdarw.
.sup.2 is a possibly nonlinear function that maps the previous state x.sub.k-1 to the current state x.sub.k with noise v.sub.k-1. The observation z.sub.k in this work is defined as the k-th X-ray image of a sequence, so that z.sub.k∈
.sup.w×h, where w and h are the width and height of an X-ray image. We further define the observation model as z.sub.k=h.sub.k(x.sub.k,n.sub.k), where n.sub.k∈
.sup.n.sup.
.sup.2×
.sup.n.sup.
.sup.w×h is a highly non-linear function that generates the observation z.sub.k from the state x.sub.k with noise n.sub.k. The state transition model f.sub.k and the observation model h.sub.k, respectively, can also be equivalently represented using probabilistic forms, i.e. the state transition prior p(x.sub.k|x.sub.k-1) and the likelihood p(z.sub.k|x.sub.k) from which x.sub.k and z.sub.k can be obtained by sampling.
(107) With these definitions and p(x.sub.0), the initial belief of x.sub.0, Bayesian filtering seeks an estimation of x.sub.k(k≥1) based on the set of all available observations z.sub.0:k={z.sub.i, 0 . . . , k} up to time k via recursively computing the posterior probability p(x.sub.k|z.sub.0:k) as Equation 4:
(108)
(109) Assuming the initial probability p(x.sub.0|z.sub.0)=p(x.sub.0) is known, based on Equation 4, Bayesian filtering runs in cycles of two steps: prediction and update. In the prediction step, the prior probability p(x.sub.k|z.sub.0:k-1), the initial belief of x.sub.k given previous observations, is estimated by computing the integral in Equation 4. In the update step, the prior probability is corrected by the current likelihood p(z.sub.k|x.sub.k) to obtain the posterior p(x.sub.k|z.sub.0:k).
(110) A Deep Learning Based Likelihood
(111) Directly modelling the likelihood p(z.sub.k|x.sub.k) is challenging due to (1) the complexity of the generation process h.sub.k and (2) the computational complexity of p(z.sub.k|x.sub.k) for every value x.sub.k∈. The problem is simplified by only computing the likelihood p(z.sub.k|x.sub.k) in the image pixel space, i.e. the integer pixel coordinate. For a subpixel x.sub.k.sub.
(112) To obtain the training labels, we assume that there exists a mapping h.sub.k, such that the training label can be defined as a distance-based probability map, i.e. the farther away x.sub.k is from the ground truth tip location in the image z.sub.k, the less possible it is to observe z.sub.k given x.sub.k through the process h.sub.k. This definition matches the intuition that from a location x.sub.k that is far from the ground truth tip location, the probability of observing a z.sub.k with the catheter tip being located at the ground truth position should be low. For simplicity, a 2D Gaussian probability density function (PDF) N(x.sub.k;x′.sub.k,σ.sup.2I) centered at the ground truth tip location x′.sub.k with variance σ.sup.2I in the image space is used as the label to train the network (1103,
(113) The network that is used follows an encoder-decoder architecture with skip connections similar to a U-net as for instance described by Ronneberger et al., “U-net: Convolutional networks for biomedical image segmentation”, International Conference on Medical image computing and computer-assisted intervention 2015, Springer. pp. 234-241. Additionally, similar to the work by Milletari et al., “V-net: Fully convolutional neural networks for volumetric medical image segmentation”, 2016 Fourth International Conference on 3D Vision (3DV), IEEE. pp. 565-571, residual blocks are adopted at each resolution level in the encoder and decoder to ease gradient propagation in a deep network. Residual blocks are basically a special case of highway networks without any gates in their skip connections. Essentially, residual blocks allows the flow of memory (or information) from initial layers to last layers as for instance described by He et al., “Deep residual learning for image recognition”, 2016 Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770-778.
(114) The encoder consists of 4 down blocks in which a residual block followed by a stride-2 convolution is used for extraction and down-scaling of feature maps. The number of feature maps is doubled in each downsampling step. The decoder has 4 up blocks where a transposed convolution of stride-2 is used for upsampling of the input feature maps. Dropout is used in the residual unit of the up block for regularization of the network. Between the encoder and the decoder, another residual block is used to process the feature maps extracted by the encoder. The detailed network architecture is shown in
(115) Due to similar appearance between a guiding catheter tip and corners of a background structure, such as vertebral bones, lung tissue, stitches or guidewires, ambiguity may exist when the network is expected to output only one blob in the probability map. To alleviate the issue, a similar strategy is adopted as taught by Laina et al., “Concurrent segmentation and localization for tracking of surgical instruments”, 2017 International conference on medical image computing and computer-assisted intervention, Springer. pp. 664-672, using a catheter mask (1102,
(116)
(117) where A is the output feature map of the 1×1 convolution, A.sub.i,j denotes the value of A at location (i, j), D is the final output of the detection network, a 2D probability map representing p(z.sub.k|x.sub.k). The details are shown in
(118) The training loss is defined as a combination of the segmentation loss and the detection loss. The segmentation loss Ls in this work is a Dice loss defined by Equation 6:
(119)
(120) where M denotes the ground truth binary catheter masks, S is the segmentation heatmap. The loss function for detection Ld is mean square error given by Equation 7:
(121)
(122) where T denotes the ground truth PDF, w and h are the width and height of an image. The total training loss L is defined as Equation 8:
L=L.sub.s+ΔL.sub.d (Equation 8)
(123) where λ is a weight to balance L.sub.s and L.sub.d.
(124) Approximation of the Posterior with Particle Filter
(125) Once the deep neural network is trained, its weights are fixed during inference for computing the posterior p(x.sub.k|z.sub.0:k) for new data. Ideally, the network detection output p(z.sub.k|x.sub.k) should be a Gaussian PDF during inference, as it is trained with labels of Gaussian PDFs. However, due to similar appearance of background structures or contrast residual, the detection output is unlikely to be a perfect Gaussian (possibly non-Gaussian or having multiple modes), which prevents the posterior p(x.sub.k|z.sub.0:k) in Equation 4 being solved with an analytical method. In practice, the posterior can be approximated using a particle filter method as for instance described by Arulampalam et al., “A tutorial on particle filters for online nonlinear/non-Gaussian Bayesian tracking”, IEEE Transactions on signal processing 2002:50, 174-188.
(126) Particle filter methods approximate the posterior PDF by a set of Ns random samples with associated weights {x.sub.k.sup.i, w.sub.k.sup.i}.sub.i=1.sup.N.sup.
P(x.sub.k|z.sub.0:k)≈Σ.sub.i=1.sup.N.sup.
(127) where δ(⋅) is the Dirac delta function. The weight w.sub.k.sup.i can be computed in a recursive manner as Equation 10 once w.sub.k-1.sup.i is known:
(128)
(129) where q(x.sub.k|x.sub.k-1.sup.i,z.sub.k) is an importance density from which it should be possible to sample x.sub.k.sup.i easily. For simplicity, a good and convenient choice of the importance density is the prior p(x.sub.k|x.sub.k-1.sup.i), so that the weight update rule (Equation 10) becomes w.sub.k.sup.i∈w.sub.k-1.sup.ip(z.sub.k|x.sub.k-1).
(130) A sample can be drawn from p(z.sub.k|x.sub.k-1.sup.i) in the following way. First, a process noise sample v.sub.k-1.sup.i is sampled from p.sub.v(v.sub.k-1), the PDF of v.sub.k-1; then x.sub.k.sup.i is generated from x.sub.k-1.sup.i via the state transition model x.sub.k.sup.i=f.sub.k(x.sub.k-1.sup.i,v.sub.k-1.sup.i). Where, p.sub.v(v.sub.k-1) is set to be a Gaussian N(0, σ.sub.v.sup.2I). The choice of motion model for f.sub.k is important for an accurate representation of the true state transition prior p(x.sub.k|x.sub.k-1). A random motion cannot characterize well the motion of catheter tip in X-ray image frames. In this application, the motion is estimated from adjacent frames using an optical flow method, as this approach 1) takes into account of the observation z.sub.k, which results in a better guess of the catheter tip motion, and 2) enables estimation of a dense motion field where the motion of a sample x.sub.k.sup.i can be efficiently obtained. Therefore, f.sub.k is defined as Equation 11:
x.sub.k=x.sub.k-1+u.sub.k-1(x.sub.k-1)+v.sub.k-1 (Equation 11)
(131) where u.sub.k-1(⋅) is the motion from frame k−1 to frame k estimated with optical flow using the method as described as for instance by Farneback et al., “Two-frame motion estimation based on polynomial expansion”, Scandinavian conference on Image analysis 2003, Springer. pp. 363-370, u.sub.k-1(x.sub.k-1)) is the motion from state x.sub.k-1.
(132) Once samples are drawn and their weights are updated, the so-called “resampling” of the samples should be performed to prevent the degeneracy problem, where all but one sample will have negligible weight after a few iterations. The resampling step resamples the existing samples according to their updated weights and then resets all sample weights to be 1/N.sub.s, so the number of effective samples which have actual contribution to approximate p(x.sub.k|z.sub.0:k) is maximized. If the resampling is applied at every time step, the particle filter becomes a sampling importance resampling (SIR) filter, and the weight update rule follows Equation 12.
w.sub.k.sup.i∝p(z.sub.k|x.sub.k.sup.i (Equation 12)
(133) The final decision on catheter tip location in frame k can then be computed as the expectation of x.sub.k,{circumflex over (x)}.sub.k=∫x.sub.k p(x.sub.k|z.sub.0:k)dx.sub.k, which is in this case, the weighted sum of all samples:
{circumflex over (x)}.sub.k=Σ.sub.i=1.sup.N.sup.
(134)
(135) The final catheter tip location as obtained from the tracking method is used in the next step 107 to determine the translation between the device location in the online fluoroscopic image and corresponding offline angiographic image of the selected roadmap.
(136) In alternate embodiments, step 106 can be accomplished by the teachings of U.S. Pat. No. 9,256,936 “Method and apparatus for tracking objects in a target area of a moving organ”. U.S. Pat. No. 9,256,936 employs the property of synchronized periodic motion to establish the dynamic geometric relationship between the positions of the tracked features. This is realized by learning the periodic motion of the individual features, such as from other visible devices generally present in the image view during percutaneous coronary interventions procedures. When one of the features is obscured, knowledge about the periodic motion patterns of both features, the position of one of the features, and the phase of the periodic motion can be used to derive the position of the obscured feature. If the first feature is the device target location itself (for instance the catheter tip), such location is immediately determined. If the first feature is not the target location, but a feature that experiences the same motion as the target location (for instance another visible devices), the position of the device target location can be derived from the position of the first feature, by using the fact that the first feature experiences the same motion as the target area. This requires knowledge about the geometric relation between the first feature and the target area.
(137) Step 107: Transform Selected Roadmap to Generate Dynamic Roadmap
(138) As described herein, a reference point can be used to compensate for motion between the roadmaps obtained from the X-ray angiographic image sequence (offline phase) with the X-ray fluoroscopic image stream (online phase). The reference point, which can be extracted from the device location, can be any object in which its motion can be correlated to the breathing motion and patient motion and the device can be for example the catheter tip, pacemaker or an anatomical landmarks. Within a preferred embodiment the device is a guiding catheter and the device location is the guiding catheter tip. Within step 107, the location of the device (e.g. catheter tip) in current X-ray fluoroscopic frame, as a result of step 106, and the device location (e.g. catheter tip) from the selected roadmap frame as a result of step 105 is used to obtain a transformation function to align the selected roadmap with the current X-ray fluoroscopic image frame. In embodiments, this transformation function might be a rigid transformation based on the displacement obtained from the catheter tip between the current frame and the catheter tip within the selected roadmap frame. Alternatively, the transformation function can be a non-rigid transformation.
(139) For example, a rigid transformation of the roadmap can be performed by using a rigid transformation function. Considering the original roadmap as function R(x,y) (as a result of step 105) and a transformation function T, then the transformation function can be as follows:
F(x,y)=T{R(x,y)} (equation 14)
(140) where F(x,y) is the transformed roadmap. The transformation function T can be, for example, a displacement function, rotation function, scaling function, etc. For instance, when the roadmap represents the vessel model as centerlines, or contours, the above transformation can be applied to each two dimensional coordinate (x,y) of the centerlines or contours. In the case where the roadmap represents the vessel model as an image mask, the above transformation can be performed on the pixels of the image mask.
(141) Step 108: Overlay Dynamic Roadmap on Fluoroscopic Image Data
(142) The dynamic roadmap that results from the cardiac matching and transformation of the roadmap is rendered and integrated as an overlay on the corresponding X-ray fluoroscopic or angiographic image data frame of the live angiographic image data image data stream of the patient.
(143) Examples of a dynamic roadmap overlaid on a non-contrast image are presented in
(144)
(145) The method provides a real time dynamic overlay or roadmapping where a visual representation of the dynamic roadmap is rendered and superimposed on the live X-ray fluoroscopic or angiographic image stream/sequence and thereby providing support for the clinician in improved patient treatment.
(146) In embodiments, step 108 can be accomplished by rendering the transformed roadmap (dynamic roadmap) that results from step 103 according to the viewpoint of the current live image in the live image stream/sequence of the patient. The pixels of the rendering of the dynamic roadmap that correspond to the organ of interest (e.g., the coronary tree) can be assigned color values in a predefined range, such as a range of color values from red to white. Besides coloring the dynamic roadmap, also transparency of the rendered model can be applied. Transparency of the model provides a dual view, where the X-ray fluoroscopy image is visible as well as the dynamic roadmap. To allow visual appreciation of stents, balloons, or other devices or instruments to treat the diseased artery and allow visual appreciation of the projected overlay, the overlay may be projected on the X-ray fluoroscopic image data in transparent mode as illustrated by 3410 in
(147) The mapping of the model (vessel layer image of
(148) The color scheme can also represent quantitative parameters. For instance geometrical parameters such as the local curvature of the vessel or the local diameter. These geometrical parameters can be derived for example from the vessel model created in step 103. Another example of quantitative parameters are pathological parameters such as the amount of calcified plaque in the lumen which is derived in step 103 and in more detail by the flowchart description of
(149) Experimental Setup of the Dynamic Coronary Roadmapping
(150) Within this section an example is provided to train the device tracking method as described by step 106 of
(151) As images may be acquired with different X-ray systems and different X-ray imaging protocols, the image dimension (number of rows and number of columns) and the pixel depth (range of pixel intensities) might vary. Therefore, all images within the datasets are resampled to a grid p×p and its intensities are scaled to a range 0 to l. Typically, p is 256 and l is 1.0.
(152) All conducted experiments in this section are outlined as follows. First a description of training the deep neural network to obtain the optimal detection is described. Next, the training (tune particle filter) of the catheter tip tracking is described. Finally, the evaluation of the trained catheter tracking is described.
(153) Training the Deep Neural Network
(154) In order to train the deep neural network to provide accurate likelihood probability map the datasets N.sub.training and N.sub.training are used. Both datasets includes labels identifying the object of interest in the datasets to be trained.
(155) Data augmentation is performed to increase the number of training samples and their diversity. The augmentation includes geometric transformation such as flipping (left-right, up-down), rotation of multiple of 90 degrees, random affine transformation (such as translation, scaling, rotation, shear), and random elastic deformation. To make the trained model robust to noise, in addition to the geometric transformations also data augmented can be performed by adding Gaussian noise to the pixel values.
(156) During training, the λ value in the training loss (equation 8) is typically set to 10 to make the scale of the two terms similar (L.sub.s and L.sub.d of equation 8). Adam optimizer (Kingsma et al., “Adam: A Method for Stochastic Optimization”, International Conference on Learning Representations, ICLR 2015) was used to minimize the loss function with a learning rate of typically 0.0001. The number of training samples in a batch is typically 4 and the network is trained with typically 100 epochs to ensure convergence.
(157) Training the deep neural network to output reasonable likelihood probability map is performed by tuning the network hyperparameters. To select hyperparameters and model weights in training, an evaluation metric is required. As the deep network is essentially a catheter tip detector, accurate detection of the tip location is desired. For example, the mean Euclidean distance between the ground truth and the predicted tip location averaged over all validation frames, can be used as the validation criteria for selecting the optimal training epoch and the network hyperparameters.
(158) Training the Catheter Tip Tracking
(159) The catheter tip is tracked in X-ray fluoroscopy images using algorithm as described by
(160) In case the approach of Farneback et al., “Two-frame motion estimation based on polynomial expansion”, Scandinavian conference on Image analysis 2003, Springer. pp. 363-370, was used as optical flow method (equation 11), a grid search to find the optimal parameter setting is done on the following parameters of the Farneback method: (1) the image scale to build the pyramids, (2) the number pyramid levels, (3) the averaging window size, (4) the number of iterations, (5) the size of the pixel neighborhood used to find polynomial expansion in each pixel, and finally (6) the standard deviation of the Gaussian that is used to smooth derivatives used as a basis for the polynomial expansion. The above parameters are tuned independently of the deep neural network, as optical flow directly estimates the catheter tip motion between two frames. To tune the parameters, we tracked the tracked catheter tip in X-ray fluoroscopy starting from the ground truth tip position in the first frame using the motion field between two adjacent frames estimated with optical flow. The average and median distance between the tracked tip position and the ground truth is than used as the evaluation criteria for the tuning.
(161) The parameters to tune for the particle filter are the number of samples Ns (See
(162) Evaluating the Catheter Tip Tracking
(163) The proposed tracking method in the algorithm described by
(164) The first one tracks catheter tip using only the detection network as described by section ‘A Deep Learning based Likelihood’ within the description of step 107 of
(165) The other two methods in this experiment use only optical flow to track catheter tip starting from the ground truth tip position in the first frame. The motion field towards the current frame, estimated by the two methods, was based on the de-formation from the previous frame or the first frame in the sequence, respectively. They are called “Optical Flow (previous)” and “Optical Flow (first)”.
(166) The tracking accuracies of all methods reported in this section were obtained on the dataset N.sub.evaluate. The tracking results of the 4 proposed methods on example image frames from the 4 X-ray fluoroscopic sequences are illustrated in
(167)
(168) Within an alternative embodiment, the set of roadmaps and dynamic roadmap generated by the process can be three-dimensional (3D) roadmaps as represented by the flowchart of
(169) Step 1801: Generate 3D Model from Data 3D Angiographic Image Data
(170) First at step 1801, which is an optional step, a 3D model of (a large portion of) the coronary tree is generated. This is done using data acquired from a 3D angiographic imaging modality, for instance computed tomography (CT), X-ray rotational angiography, 3D Ultrasound, or magnetic resonance imaging (MRI). The 3D model can be for example in the form of 3D centerlines, 3D surface contours representing the luminal surface of the vessels and/or the outer vessel surface, plaque, 3D masks, or a combination of these.
(171) The 3D centerlines can be created manually, for instance by indicating the vessel centerlines within the 3D volumetric image data, or automatically as taught for instance by Lesage et al, “Bayesian Maximal Paths for Coronary Artery Segmentation from 3D CT Angiograms”, MICCAI 2009, Part 1, LNCS 5761, pp 222-229.
(172) The coronary lumen, arterial wall and detection of coronary plaque from CT angiographic image data can be for (semi) automatically detected as for instance taught by Kirissli et al. “Standardized evaluation framework for evaluating coronary artery stenosis detection, stenosis quantification and lumen segmentation algorithms in computed tomography angiography”, Medical Image Analysis, vol. 17, no. 8, pp. 859-876, 2013, a methods are described to detect the coronary lumen, arterial wall and detection of coronary plaque from CT angiographic image data.
(173) Coronary plaque can be detected as for instance disclosed in U.S. application Ser. No. 16/379,248 (Method and system for assessing vessel obstruction based on machine learning) or by for instance the method as taught by Dey et al., “Automated 3-dimensional quantification of noncalcified and calcified coronary plaque from coronary CT angiography”, Cardiovasc Comput Tomogr. 2009; 3(6):372-382, or as taught by Gerretsen et al., “Visualization of coronary wall atherosclerosis in asymptomatic subjects and patients with coronary artery disease using magnetic resonance imaging”, PLoS One. 2010 Sep. 29; 5(9), or as taught by Adame et al., “Automatic segmentation and plaque characterization in atherosclerotic carotid artery MR images”, Magnetic Resonance Materials in Physics, Biology and Medicine 2004; 16 (5): 227-234.
(174)
(175) Optionally, the 3D angiographic CT image data can be used to compute the cFFR along the coronary tree as for instance disclosed by U.S. application Ser. No. 16/379,248, which discloses methods to compute the fraction flow reserve along a coronary artery. Optionally, the extracted 3D model either from CT, rotation angiography or MRI can be used to compute the coronary WSS, or time average WSS as for instance taught by Wentzel et al., “Geometry guided data averaging enables the interpretation of shear stress related plaque development in human coronary arteries”, Journal of Biomechanics 2005, 1551-1555.
(176) Step 1802: Retrieve X-Ray Angiographic Image Data
(177) Within step 1802, the X-ray angiographic image data is retrieved and is similar to step 101 of
(178) For the creation of the 3D roadmap as described by step 1803, it is important that the X-ray angiographic image sequence(s) are taken from a right perspective which is defined as the angulations of an X-ray system (both the system rotation and angulation) that contains as much information regarding the segment of interest as possible. In this perspective foreshortening and overlap of surrounding vessels are minimized. Foreshortening is the event when an object seems compressed when viewed from a certain perspective, causing distortion in the information. The perspectives in which an object of interest are visualized with minimum foreshortening are called optimal perspectives. In a preferred embodiment the 3D model as obtained from the 3D angiographic image data modality as described by step 1801 is used to suggest optimal perspectives to the user in terms of foreshortening and overlap of surrounding vessels. Of this 3D model the orientation is known relative to the X-ray system. An optimal perspective in terms of minimal foreshortening is determined as a perspective that is perpendicular to the orientation of the 3D model or a section thereof. Because the model can be looked at from different angles that are all perpendicular to it, a various number of optimal perspectives is possible.
(179) However, an optimal perspective is not solely dependent on minimizing foreshortening but also on overlap of surrounding vessels. Therefore, a measure for this overlap is also taken into account. The overlap of surrounding vessels can be for one or multiple heart phases because due to movement, either of the heart itself or breathing motion, surrounding vessels can overlap the segment of interest during a certain time moment.
(180) The 3D model extracted model as a result of step 1801 is back projected onto a 2D plane representing a certain perspective as taught by Lay, “Linear algebra and its applications”, 2012, 4th edition, p142-143, Addison-Wesley Longman. Every point that is within the vessel of interest in the 3D model, is assigned a certain value. For every 3D point, its value is added to the corresponding 2D point in the back projected image. The plane with the maximum amount of 2D points containing a value, is the most desirable perspective in terms of minimal overlap.
(181) Additionally, a perspective can be indicated in terms of minimal in-plane coronary motion. This perspective shows the vessel of interest with the least amount of in-plane coronary motion in the perspective. This allows the clinician to view the vessel of interest in a position where the vessel is as still as possible. A measure for the amount of in-plane coronary motion for each perspective can be determined for instance by back projecting the 3D model extracted from the CT data onto a 2D plane representing a certain perspective as taught by Lay, “Linear algebra and its applications”, 2012, 4th edition, p142-143, Addison-Wesley Longman. For every centerline point of each vessel in the 3D model, a position is known. Then the 3D model extracted from CT data can be deformed using the motion model as taught by Baka et al, “3D+t/2D+t CTA-XA registration using population-based motion estimates”, Demirci, Lee, Radeva, Unal (eds): MICCAI-STENT 2012, pp 64-71 to yield a 3D model at a different time moment. This deformed 3D model is then also back projected onto a 2D plane representing the certain perspective. Again the position of each centerline point is known but now for a different moment in time. For every centerline point the in-plane coronary motion can be determined by comparing the positions of each centerline point in both back-projections. The plane with the minimum amount of in-plane movement for all centerline points, is the most desirable perspective in terms of minimal in-plane coronary motion. This can be done for one or multiple heart phases.
(182) Then for every combination of angulation and rotation (thus each perspective), it can be indicated how optimal the resulting perspective is. This indication is for instance a weighted sum of foreshortening, overlap of surrounding vessels and/or in-plane coronary motion for at least one heart phase. When multiple heart phases are taken into account, the calculations are done for each time moment that is for each frame. A weighted sum of all frames within the multiple heart phases is then made to obtain an overall indication of how optimal the perspectives are. This can be visualized for instance using a color map as shown in
(183) In case the 3D model is not available, two X-ray angiographic image sequence are required. The first X-ray angiographic image sequence is similar as described by step 101 of
(184) Step 1803: Create 3D Roadmaps
(185) Within this step the sequence of 3D roadmaps (3D+t) will be created and is similar to step 103 of
(186) The first method (1803a) to create the 3D+t roadmaps is applicable in case step 1801 is not performed, and the 3D+t roadmaps are created by processing the two X-ray angiographic image sequences that result from step 1802. The 3D+t roadmap for all frames within one cardiac cycle after contrast injection can for instance be created by the method as taught by Chen et al, “Kinematic and Deformation Analysis of 4-D Coronary Arterial Trees Reconstructed From Cine Angiograms”, IEEE Transactions on medical imaging, Vol. 22, No. 6, June 2003 pp 710-721, or as taught by Zheng et al, “Sequential reconstruction of vessel skeletons from X-ray coronary angiographic sequences”, Computerized Medical Imaging and Graphics 34 (2010) 333-345. An example of a single 3D roadmap frame as a result of this step is provided by 2402 of
(187) The second method (1803b) to create the 3D+t roadmaps is applicable when the 3D model is available as a result from step 1801 and one X-ray angiographic image sequence is available as a result from step 1802. Based on a method as taught by for instance Baka et al, “Statistical coronary motion models for 2D+t/3D registration of X-ray coronary angiography and CTA”, Medical Image Analysis, Volume 17, Issue 6, August 2013, Pages 698-709. Within this work Baka et al proposed a method for building population based average and predicted motion from 4D CT datasets which is then used to perform 3D+t/2D+t registration based on distance minimization on one cardiac cycle. An example of a single 3D roadmap frame as a result of this step is provided by 2408 of
(188) The third method (1803c) to create the 3D+t roadmaps is applicable when the 3D model is available as a result from step 1801 and two X-ray angiographic image sequence are available as a result from step 1802. In this method both the 3D model resulting from step 1801 and a 3D reconstruction based on the X-ray angiographic image data as a result of step 1802 is combined to create the 3D+t roadmaps. This can be achieved as for instance by the methodology described in U.S. Pat. No. 10,229,516 “Method and Apparatus to Improve a 3D+Time Reconstruction”, which describes a method for making a three-dimensional surface reconstruction over time of an object from two or more bi-dimensional x-ray images of the object. Alternatively, the 3D+t roadmaps can be created as for instance taught by Dibildox et al., “3D/3D registration of coronary CTA and biplane XA reconstructions for improved image guidance”, Med Phys. 2014 Sep; 41(9), or as taught by Baka et al., “Oriented Gaussian mixture models for nonrigid 2D/3D coronary artery registration”, IEEE Trans Med Imaging. 2014 May; 33(5):1023-34. An example of a single 3D roadmap frame as a result of this step is provided by 2410 of
(189) Optionally, based on the generated roadmaps, quantitative image analysis can be performed to extract clinical relevant information, such as for example location and percentage of vessel obstruction, diameter and area, length of the vessel or the curvature of the vessel as for instance taught by Girasis et al. in “Advanced three-dimensional quantitative coronary angiographic assessment of bifurcation lesions: methodology and phantom validation”, EuroIntervention. 2013 Apr. 22; 8(12):1451-60 or as taught by Wang et al., “Vessel extraction in coronary X-ray Angiography”, Conf Proc IEEE Eng Med Biol Soc. 2005; 2: 1584-1587.
(190) Optionally, the generated roadmaps can be used to compute the cFFR along the coronary tree as for instance disclosed by U.S. application Ser. No. 16/438,955, which discloses methods to compute the fraction flow reserve along a coronary artery based on 3D coronary reconstruction obtained with X-ray angiography. The 3D+t model enables the calculation of the cFFR for every time point and takes into account the geometrical variation of the vessel tree part during a cardiac cycle and hemodynamic variation during the cardiac cycle. Optionally, the generated roadmaps can be used to compute the coronary WSS, or time average WSS as for instance taught by Wentzel et al., “Geometry guided data averaging enables the interpretation of shear stress related plaque development in human coronary arteries”, Journal of Biomechanics 2005, 1551-1555. The time averaged WSS can be also be derived by calculating the WSS for all time points t within the 3D+t model.
(191) Optionally, the location and amount of calcified plaque can be extracted from the X-ray angiographic image sequence(s) as for instance disclosed in detail by the flowchart description of
(192) Step 1804: Obtain Device within X-Ray Angiographic Image Data
(193) Within step 1804 information is obtained to allow alignment (step 1808) of the 3D+t roadmap (as a result from step 1803) during the live overlay (1809) as part of the online phase (1810), and step 1804 is similar to step 102 of
(194) In case step 1801 is not performed, this step (1804) is identical to the steps 501, 502 and 503 as described by
(195) In case the 3D+t roadmap is created by the method 1803b, steps 501, 502 and 503 are applied to the single X-ray angiographic image sequence and no 3D reconstruction will be performed of the device. However, the device location as a result from step 503 is integrated in the 3D+t roadmap as created by the method 1803b. An example of a single 3D roadmap frame as a result of this sub-step is provided by 2405 of
(196) In case the 3D+t roadmap is created by the method 1803c as described by step 1803 and at least two x-ray angiographic image sequences are available as a result from step 1802, the above described method is applicable as well. An example of a single 3D roadmap frame as a result of this sub-step is provided by 2409 of
(197) Step 1805: Retrieve Fluoroscopic Image Data
(198) This step is identical to step 104 of
(199) Optionally, since arbitrary X-ray system geometry is allowed within the flowchart of
(200) Step 1806: Select 3D Roadmap
(201) This step is identical to step 105 of
(202) Alternatively, in the case that the X-ray fluoroscopy image data is obtained with a bi-plane acquisition, the images from both projections can be acquired shortly after each other with a small time delay, resulting in a high temporal resolution. Due to this high temporal resolution a more accurate roadmap selection takes place.
(203) Step 1807: Track Device
(204) This step is identical to step 106 in
(205) Alternatively, in the case that the X-ray fluoroscopy data is obtained with a bi-plane acquisition, the tracking of the device can be performed in both X-ray fluoroscopy image projections separately according to the technique described in step 106 of
(206) Step 1808: Transform Selected 3D Roadmap to Generate Dynamic 2D Roadmap
(207) This step is similar to step 107 in
(208) Within step 1808, the location of the device in current X-ray fluoroscopic frame, as a result of step 1807, and the device location (e.g. catheter tip) from the selected 3D roadmap frame as a result of step 1804 is used to obtain a transformation function to align the selected roadmap with the current X-ray fluoroscopic image frame. In embodiments, this transformation function might be a rigid transformation based on the displacement obtained from the catheter tip between the current frame and the catheter tip within the selected roadmap frame. Alternative the transformation function is a non-rigid transformation.
(209) For example, a rigid transformation of the roadmap can be performed by using a rigid transformation function. Considering the original 3D roadmap as function R(x,y,z) (as a result of step 1806) and a transformation function T, then the transformation function can be as follows:
F(x,y,z)=T{R(x,y,z)} (equation 15)
(210) where F(x,y,z) is the transformed 3D roadmap. The transformation function T can be for example a displacement function, rotation function, scaling function, etc. Since arbitrary X-ray system geometry is allowed during the acquisition of the X-ray fluoroscopic frames by step 1805, the transformation function can take into account X-ray angulation, magnification and table movement and creates a 2D roadmap representing an imaging plane of the current X-ray fluoroscopic image frame. For instance, when the roadmap represents the vessel model as centerlines, or contours, the above transformation can be applied to each three-dimensional coordinate (x,y,z) of the centerlines or contours. In the case where the roadmap represents the vessel model as an image mask, the above transformation can be performed on the voxels (which map to three-dimensional coordinates x,y,z) of the volumetric image mask.
(211) Alternatively, in the case that the X-ray fluoroscopy data is obtained with a bi-plane acquisition, the tracking of the device can be performed in both X-ray fluoroscopy image projections. In this case, the transformation as described above can be performed for each fluoroscopic projection and two 2D dynamic roadmaps are created.
(212) Step 1809: Overlay Dynamic Roadmap on Fluoroscopic Image Data
(213) This step is similar to step 108 in
(214) In embodiments, step 1809 can be accomplished by rendering the transformed model that results from step 1808 according to the viewpoint of the current live image in the live image stream/sequence of the patient. The pixels of the rendering of the model that correspond to the organ of interest (e.g., the coronary tree) can be assigned color values in a predefined range, such as a range of color values from red to white. Besides coloring the model, also transparency of the rendered model can be applied. Transparency of the model provides a dual view, where the X-ray fluoroscopy image is visible and the
(215) The mapping of the model to the dynamic roadmap overlay can be performed by translating the pixel intensities to color values where the color scheme represent quantitative parameters as described within step 1804.
(216) Additionally, multiple overlays are rendered and integrated as an overlay on the live X-ray fluoroscopic or angiographic image stream. The overlay can exist of the 3D roadmap and in addition to the roadmap for example a 3D volumetric data or quantitative parameters. The 3D volumetric data needs to be back projected on the current X-ray fluoroscopic image frame based on the projection (view angle) in the same way as the 3D roadmap.
(217) Obviously, step 1808 and 1809 might be performed in a slightly different order, in which in step 1808, results in a 3D transformed roadmap based on the information from step 1808 and 1807, and that in step 1809, this transformed 3D roadmap is back projected onto the current X-ray fluoroscopic image frame in which step 1809 handles the current X-ray fluoroscopic angulation, magnification and table movement during the back projection.
(218) After step 1809, the method continues with step 1805 to process the next X-ray fluoroscopic image frame(s) and in case the retrieval of the X-ray fluoroscopic image data stops, as described by step 1805, the flowchart as described by
(219) Within an alternative embodiment, subtraction is used as an image enhancement technique to improve the identification and processing of the object of interest. Image enhancement is a useful tool to improve the identification of an object of interest. An example of image enhancement is subtraction, where static objects and background are removed from the image of interest by subtracting this image with a baseline image.
(220)
(221) First the X-ray angiographic image data is retrieved 1501 including ECG, identical to 101 in
(222) Next in step 1503, one whole cardiac cycle of the X-ray fluoroscopic image data of 1502 is selected. For every frame of the X-ray angiographic image data of 1501 a baseline image frame of 1502 is selected based on ECG matching. The ECG matching can be performed as described in 105 of
(223) The enhanced image resulting from 1503 is used to obtain the device location in step 1504. Step 1504 is identical to step 102 in
(224) In the online phase 1512, the fluoroscopic image data is retrieved 1506 including ECG, identical to 104 with the same projection (view angle) as 1501 and 1502. Next in step 1507, image enhancement is applied to the x-ray fluoroscopic image of 1506, and the same process as in step 1503 is executed. This includes ECG matching of the X-ray fluoroscopic image data 1506 and the baseline data 1502. Subtraction of the X-ray fluoroscopic image data 1506 and the matched the baseline data 1502, resulting in enhanced x-ray fluoroscopic image data. Additionally, image processing techniques such as edge enhancement can be applied to improve the outline the object of interest.
(225) Subsequently in step 1508, roadmap selection takes place identical to step 105 in
(226) Within an alternative embodiment, the online phase of the dynamic roadmapping can be reinitiated by means of an X-ray angiographic image stream and is represented by the flowchart of
(227) The flowchart of
(228) Step 2601: Offline Phase
(229) Within step 2601, the sequence of roadmaps are created, and the steps are identical to the offline phase (100) represented by step 101, 102 and 103 of
(230) Step 2602: Initiate on X-Ray Angiography
(231) Within this step, the roadmap is realigned to the current situation. With the term current situation refers to a true live status during the online phase (2610). Such a realignment may be useful for instance after table movement during the online phase (2610) or any other situation in which it plausible that the roadmap is misaligned to the current situation, or when additional devices which are introduced during the online phase to allow more accurate alignment. This step may be omitted when the system make the first transition from step 2601 to the online phase (2610). Step 2602 may be initiated from a signal triggered by the X-ray system, or another external system, after a contrast bolus injection was initiated. Alternatively, a contrast bolus injections is automatically detected by analyzing the X-ray fluoroscopic image stream (2603) as for instance taught by Ma et al., “Fast prospective detection of contrast inflow in x-ray angiograms with convolutional neural network and recurrent neural network”, International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer (2017) 453-461, or as described by step 501 of
(232) The first sub-step within 2602 is the retrieval an X-ray angiographic image stream, which can be either obtained from a single plane or bi-plane acquisition. Once an amount of X-ray angiographic image frames are available, preferable an amount of frames covering at least one cardiac cycle, the roadmap is realigned to the current situation. Simultaneous with the X-ray angiographic image stream, the ECG signal is obtained as well.
(233) In case the offline phase represents a 2D roadmap, the second sub-step within 2602 deals with update (recreation, or realignment) of the 2D roadmap sequence, and is identical to step 102 and 103 of
(234) Alternatively, in the case that the image data is retrieved from acquiring multiple mono-plane images with different projections or a bi-plane acquisition in the offline phase, the device location can be determined according to step 102 in
(235) Step 2603: Retrieve X-Ray Fluoroscopic Image Data
(236) Within step 2601, the X-ray fluoroscopic image data is retrieved, and this step is identical to step 104 of
(237) Step 2604: Select Roadmap
(238) Within step 2604 the current roadmap from the sequence of roadmaps as a result of step 2601 or in case updated, as a result from step 206 is selected. When the offline phase represents a 2D roadmap, this step is identical to step 105 of
(239) Alternatively, the X-ray fluoroscopy data is obtained with a bi-plane acquisition. With a bi-plane system the images from both projections are acquired shortly after each other with a small time delay, resulting in a high temporal resolution. Due to this high temporal resolution a more accurate roadmap selection takes place.
(240) Step 2605: Track Device
(241) Inn step 2605 the device is tracked within the live X-ray fluoroscopic image steam. When the offline phase represent a 2D roadmap, this step is identical to step 106 of
(242) Alternatively, the X-ray fluoroscopy data can be obtained with a bi-plane acquisition. In this case, the tracking of the device can be performed in both X-ray fluoroscopy image projections separately according to the technique described in step 106 of
(243) Next to that, the device location can be determined as a 3D location based on both projections from the bi-plane acquisition as for instance taught by Ambrosini et al., “A Hidden Markov Model for 3D Catheter Tip Tracking With 2D X-ray Catheterization Sequence and 3D Rotational Angiography”, IEEE Trans Med Imaging. 2017 March; 36(3):757-768. This enables a 3D tracking of the device.
(244) Step 2606 Transform Selected Roadmap to Generate Dynamic Roadmap
(245) In step 2606 the selected roadmap is transformed to create a dynamic roadmap for the current live X-ray fluoroscopic image frame. In the case that the offline phase represents a 2D roadmap, this step is identical to step 107 of
(246) Alternatively, the X-ray fluoroscopy data can be obtained with a bi-plane acquisition. In this case, two translations are obtained in step 2605. These two translations and the 3D orientation of both image projections, results in a 3D transformation of the selected roadmap.
(247) Step 2607: Overlay Dynamic Roadmap on X-Ray Fluoroscopic Image Data
(248) When the offline phase produces 2D roadmaps, this step is identical to step 107 of
(249) Step 2608: Detect Drift
(250) During step 2608, the system detects if a realignment of the roadmap is required. For instance the X-ray systems triggers a signal incase the table is adjusted during the online phase (2610) or after a contrast administration was initiated. Optionally, within step 2608, misalignment of the roadmap can be detected by image processing of the X-ray fluoroscopic image stream. For instance comparison of the current image frame with the image frame at the same cardiac phase of the previous cardiac cycle. The comparison between both images can be performed by for example image registration, cross correlation or (minimum) difference. When applying image registration, the displacement resulting from the image registration is a measure of variation between both images and a threshold value can be defined to detect drift. In case of cross correlation, the cross correlation value represents the similarity between both images and for example a threshold value can be defined to detect drift. Alternatively, one image is shifted with respect to the other image and the cross correlation between both images is calculated for multiple shifts. Next, the maximum cross correlation value might correspond with a certain shift of the image. The magnitude of the shift is a magnitude of the drift. In case of the difference between both images, the difference value is a measure for the drift. A threshold value might be defined for drift detection. Alternatively, one image is shifted with respect to the other image and the difference between both is images calculated. The minimum difference can be calculated for multiple shifts and the shift that corresponds with the minimum difference is a measure for the drift.
(251) Another method for drift detection is evaluation of the ECG signal. The ECG is a cyclic signal. In case of for example arrhythmia, the heartbeat is irregular and causes abnormal cardiac motion and the ECG signal deviates from a normal cycle. Deviations in the ECG signal may hamper the roadmap selection and may thereby cause incorrect overlaying of the roadmap. Therefore, evaluation of the ECG signal is useful to detect drift.
(252) During regular respiration the diaphragm contracts, so called eupnea. For example, in case the patient suffers from hiccups, the diaphragm motion is irregular and therefore the motion of the object of interest might be irregular as well. The diaphragm contraction might be visible in the X-ray angiographic images. Tracking of the diaphragm in the X-ray image sequence can help to detect irregular respiration and the necessity for drift correction. Tracking the motion of the diaphragm can be performed as for instance taught by Ma et al., “PCA-derived respiratory motion surrogates from X-ray angiograms for percutaneous coronary interventions”, Int J Comput Assist Radiol Surg. 2015 June; 10(6):695-705. Therefore, evaluation of the diaphragm contraction (motion) is useful to detect drift.
(253) In case that step 2610 indicates that misalignment is present, step 2602 is initiated, otherwise the system continues with step 2603. In case the retrieval of the X-ray fluoroscopic image data as described by step 2603 stops, the flowchart described by
(254) Although calcified plaque is radiopaque, its presents cannot be appreciated on a single X-ray image frame, and is hardly visible when assessing a sequence of X-ray images.
(255) First at step 2801, the X-ray image data is retrieved. As described before this can be either an X-ray fluoroscopic image sequence or an X-ray angiographic image sequence which contains image data before and after the administration of a contrast agent. First the steps will be described in case the X-ray image data represents X-ray angiographic image data, and afterwards the method is explained in case the X-ray image data represents X-ray fluoroscopic image data. In all cases it is assumed that the ECG signal is part of the X-ray image data.
(256)
(257) Alternatively, the properties of the X-ray beam may be changes in such a way that calcified tissue absorbs X-ray more efficient than its surrounding tissue. In applications for assessment of vascular structures, the properties of the X-ray beams are defined in such a way that there is an optimal X-ray absorption of the iodine rich contrast liquid with respect to its surrounding tissues. There are two primary means that define the properties of the X-ray beam produced by the tube: a) altering the current (mA) and b) altering the voltage (kV). The current (measured in and often referred to as mA, or milliamperes) across the tube determines how many electrons are released to strike the anode. Increasing the mA will increase the number of electrons that strike the anode, with a consequent linear increase in the number of photons produced by the tube. The voltage across the X-ray tube (measured in and often referred to as kV, or kilovolts) affects the velocity of the electrons as they strike the anode; this affects the energy of the photons that can be produced by the tube. Additionally, higher velocity electrons will produce more photons, something on the order of (kV).sup.3. There are two major ways in which X-ray beams interact with tissue. The first is the photoelectric effect, where a photon uses up all of its energy to eject an electron from an atom; while the electron will move around and ionize neighboring atoms, there are no scatter photons. The second major effect is Compton (incoherent) scatter, where a photon hits an atom and ionizes an electron but does not use up all of its energy. The photon then scatters in a different direction with a bit less energy, and the free electron goes about doing damage. Scattered photons can travel back towards the tube, pass through the patient and hit the detector from any odd angle, or scatter again within the patient. As the X-ray beam passes through tissue, photons get absorbed so there is less energy; this is known as attenuation. Higher energetic photons travel through tissue more easily than lower energetic photons (this means that the higher energy photons are less likely to interact with matter). Much of this effect is related to the photoelectric effect; the probability of photoelectric absorption is approximately proportional to (Z/E).sup.3, where Z is the atomic number of the tissue atom and E is the photon energy. As E gets larger, the likelihood of interaction drops rapidly. Compton scattering is about constant for different energies although it slowly decreases at higher energies. So alternatively, the X-ray fluoroscopic image sequence or the X-ray angiographic image sequence is acquired in which the tube properties (tube voltage and/or tube current) of the X-ray system is optimized in such a way that calcified tissue absorbs X-ray more efficient than its surrounding tissue.
(258) Next at step 2802, the vessel of interest is obtained. This can be either by methods as described by step 103 of
(259) At step 2803 the frame from the X-ray angiographic image sequence in which the contrast liquid enters the coronary arteries is identified and identified as f.sub.start-contrast. This step can be accomplished for instance by the methods as described by step 501 of
(260) Next, at step 2804 the vessel skeleton is detected in each frame within the sequence after the start of contrast (f.sub.start-contrast) as a result of step 2803, and preferably the vessel skeleton is detected in an amount of sequential frames representing at least one cardiac cycle. Preferable at least three cardiac cycles are available. The vessel skeleton can be detected for instance as taught by Zheng et al., “Sequential reconstruction of vessel skeletons from X-ray coronary angiographic sequences”, Computerized Medical Imaging and Graphics 34 (2010) 333-345. An example of is such a vessel detection within two X-ray angiographic image frames is provided by 3001 and 3002 within
(261) Within step 2806, the selected corresponding fluoroscopic images as a result of step 2805 are registered to each other. The registration is performed on each decomposed vessel branch as described by step 2804. Optionally, landmarks are automatically identified from the skeleton and/or decomposed skeleton to improve the performance of the registration, such landmarks can for instance be the point of bifurcation (3104) or start/end position (3105) or high local curvature (3106). The 2D to 2D registration can be performed for instance as taught by Maintz et al., “An overview of Medical Image registration Methods”, In symposium of the Belgian hospital physicists association, 1996, or as taught by Rohr at al., “Landmark-based elastic registration using approximating thin-plate splines”, IEEE Trans Med Imaging. 2001 June; 20(6):526-34.
(262) Within step 2807, the image in which the calcified plaque is enhanced is generated. Before generating the enhanced image, the registered X-ray fluoroscopic images (as a result of step 2806) are preprocessed. X-ray images typically suffer from a transparent background layer potentially obscuring the calcified plaque. In order to enclose the influence of this effect, background subtraction is performed. A simple and effective method for removing a static transparent layer uses the assumption that each individual layer can only add mass. Taking for each pixel the maximum intensity over time (lower pixel intensities absorbs x-ray) will yield an image showing the least amount of mass over time. Assuming that all pixels will not contain any contributions from moving mass layers at some point in time, the maximum intensity image is equal to the static background layer. Often, a single general mask image is created for each frame, for instance based on the maximum intensity of each pixel through all frames like discussed before. However, such a single mask image suffers from artifacts due to large, slow moving objects in the background (like for instance the diaphragm, ribs and/or lungs). Therefore, the background subtraction performed determines a more local background mask of x amount of successive frames symmetrically around the respective frame, in which x is typically 5 frames. Alternatively, the static background layer can be generated as taught by Ma et al., “Layer separation for vessel enhancement in interventional X-ray angiograms using morphological filtering and robust PCA”, Workshop on Augmented Environments for Computer-Assisted Interventions 2017, Springer. pp. 104-113 or by Ma et al., “Automatic online layer separation for vessel enhancement in X-ray angiograms for percutaneous coronary interventions”, Med Image Anal. 2017 July; 39:145-161.
(263) Next, the preprocessed frames are combined into a single image frame. This is performed by adding (3008 of
(264) Before continuing with step 2808, the method of
(265) Referring back to
(266) In general, the X-ray image data contains corresponding ECG signal recording. In case the X-ray image data does not contain ECG information, cardiac phase matching of the contrast and non-contrast images based on the ECG signal is not possible. To overcome this problem, a heart model can be used that mimics the cardiac motion. Such a heart model might be generated from for example CT acquisitions of a number of hearts. The cardiac motion is extracted from the image data and a model is generated that mimics the cardiac motion.
(267) The cardiac motion model provides the cardiac motion in 3D+t. Based on the X-ray angiography acquisition, the expected cardiac motion of the heart in a specific projection can be extracted from the cardiac motion model.
(268) The cardiac motion model is applied to both the contrast and non-contrast image and for example the correlation between both images is calculated to identify the best match between both images as for instance taught by Metz et al., “Patient Specific 4D Coronary Models from ECG-gated CTA Data for Intra-operative Dynamic Alignment of CTA with X-ray Images”, Med Image Comput Assist Interv. 2009; 12(Pt 1):369-76.
(269) Other Applications
(270) The embodiments described above are associated to provide a real time dynamic overlay or dynamic coronary roadmap which can be superimposed on the live X-ray fluoroscopic or angiographic image stream/sequence.
(271) In embodiments, the methods described within the current application can also be used to provide static guidance, meaning selecting of only one roadmap within the sequence of roadmaps. Preferably, a roadmap is selected in which the coronary artery has the least cardiac motion. Furthermore, the 3D model as created from 3D angiographic image data as described by step 1801 of
(272) In other embodiments, the described method can be used as pre-procedural planning, and optionally X-ray angiographic image can be simulated as for instance described by U.S. Pat. No. 10,192,352 for improved preparation for the actual PCI procedure.
(273) Alternatively, a single roadmap can be selected. The image related to the selected roadmap can be used as reference image. The relation between the roadmap and the reference image can be determined, for example, using the ECG signal of the patient. Next, all images of the live X-ray fluoroscopic or angiographic image stream/sequence are registered with the reference image. Due to the image registration, the images of the live X-ray fluoroscopic or angiographic image stream/sequence are “freezed” and are aligned with the static roadmap. The image registration can be performed by methods known in the art, which are taught by for example Maintz et al., “An overview of Medical Image registration Methods”, In symposium of the Belgian hospital physicists association, 1996.
(274) The embodiments described above are associated to provide a real time dynamic overlay or dynamic coronary roadmap which can be superimposed on the live X-ray fluoroscopic or angiographic image stream/sequence. Alternatively, the embodiments, with special focus on the extract roadmaps as described by
(275) From the first introduction of PCI in the 1970s, interventional cardiology has undergone significant evolution in device technology, treatment procedures and pharmacotherapy, which have mean treatment of the most complex lesion (such as total coronary occlusion) possible. Despite the steady growth of progress in nearly all facets of the coronary intervention field, the mechanical aspects of PCI, such as manipulating of coronary guidewires, balloons and stents and the occupational hazards for operators and catheterization laboratory staff remains largely unchanged since its introduction in the 1970s. The interventional cardiologist works under the guidance of direct fluoroscopy to manipulate intravascular devices, and this requires dress in heavy protective garments. Over the course of an interventional cardiology career, operators are subject to the adverse consequences of cumulative radiation exposure and an increased prevalence of orthopedic injuries.
(276) A robotic system that takes over the manipulations of the intravascular devices would significantly decrease the above mentioned short comes and hazards and could revolutionize percutaneous coronary intervention procedures. Such a robotic assisted PCI system, although still in its infancy, is for instance manufactured by Corindus vascular robotics.
(277) One of the requirements of such a robotic system is to manipulate at submillimeter level the intravascular device and requires knowledge of the 3D geometry and morphology of the vasculature. This also involves knowledge during the procedure on changes in 3D shape of the vasculature as for instance due to cardiac motion, breathing motion and/or patient motion. Using the methods as described in current application integrated in a robotic system that takes over the manipulations of the intravascular devices can potentially improve patient outcomes and allows an advanced pre-procedure planning tool with robotic precision to bring a new revolutionary standard of care to patients.
(278) Operations can be performed by processor unit on a standalone system, or a semi-standalone system which is connected to the X-ray system (
(279) Portions of the system (as defined by various functional blocks) may be implemented with dedicated hardware, analog and/or digital circuitry, and/or one or more processors operating program instructions stored in memory.
(280) The X-ray system of
(281) An X-ray beam 1703 comprises of photons with a spectrum of energies that range up to a maximum determined by among others the voltage and current submitted to the X-ray tube 1701. The X-ray beam 1703 then passes through the patient 1704 that lies on an adjustable table 1705. The X-ray photons of the X-ray beam 1703 penetrate the tissue of the patient to a varying degree. Different structures in the patient 1704 absorb different fractions of the radiation, modulating the beam intensity. The modulated X-ray beam 1703′ that exits from the patient 1704 is detected by the image detector 1706 that is located opposite of the X-ray tube. This image detector 1706 can either be an indirect or a direct detection system.
(282) In case of an indirect detection system, the image detector 1706 comprises of a vacuum tube (the X-ray image intensifier) that converts the X-ray exit beam 1703′ into an amplified visible light image. This amplified visible light image is then transmitted to a visible light image receptor such as a digital video camera for image display and recording. This results in a digital image signal. In case of a direct detection system, the image detector 1706 comprises of a flat panel detector. The flat panel detector directly converts the X-ray exit beam 1703′ into a digital image signal. The digital image signal resulting from the image detector 1706 is passed through a digital image processing unit 1707. The digital image processing unit 1707 converts the digital image signal from 1706 into a corrected X-ray image (for instance inverted and/or contrast enhanced) in a standard image file format for instance DICOM. The corrected X-ray image can then be stored on a hard drive 1708.
(283) Furthermore the X-ray system of
(284) The X-ray system of
(285) Additionally, the adjustable table 1705 can be moved using the table control 1711. The adjustable table 1705 can be moved along the x, y and z axis as well as tilted around a certain point.
(286) Furthermore a measuring unit 1713 is present in the X-ray system. This measuring unit contains information regarding the patient, for instance information regarding ECG, aortic pressure, biomarkers, and/or height, length etc.
(287) A general unit 1712 is also present in the X-ray system. This general unit 1712 can be used to interact with the C-arm control 1710, the table control 1711, the digital image processing unit 1707, and the measuring unit 1713.
(288) An embodiment is implemented by the X-ray system of
(289) The X-ray image sequences are then generated using the high voltage generator 1702, the X-ray tube 1701, the image detector 1706 and the digital image processing unit 1707 as described above. These images are then stored on the hard drive 1708. Using these X-ray image sequences, the general processing unit 1712 performs the methods as described by present application, as for instance as described by
(290) As mentioned before the operations can also be performed by a processor unit of a semi-standalone system which is connected to the X-ray system (
(291) There have been described and illustrated herein several embodiments of a method and apparatus for quantitative flow analysis. While particular embodiments have been described, it is not intended that the invention be limited thereto, as it is intended that the invention be as broad in scope as the art will allow and that the specification be read likewise. For example, the data processing operations can be performed offline on images stored in digital storage. This is typically done in a universal language (vendor independent) such as DICOM (Digital Imaging and Communications in Medicine). The storage can be a hard disk or a PACS (picture archiving and communications system) server or a VNA (vendor neutral archive) or other picture archiving and communication systems commonly used in the medical imaging arts. It will therefore be appreciated by those skilled in the art that yet other modifications could be made to the provided invention without deviating from its spirit and scope as claimed.
(292) The embodiments described herein may include a variety of data stores and other memory and storage media as discussed above. These can reside in a variety of locations, such as on a storage medium local to (and/or resident in) one or more of the computers or remote from any or all of the computers across the network. In a particular set of embodiments, the information may reside in a storage-area network (“SAN”) familiar to those skilled in the art. Similarly, any necessary files for performing the functions attributed to the computers, servers or other network devices may be stored locally and/or remotely, as appropriate. Where a system includes computerized devices, each such device can include hardware elements that may be electrically coupled via a bus, the elements including, for example, at least one central processing unit (“CPU” or “processor”), at least one input device (e.g., a mouse, keyboard, controller, touch screen or keypad) and at least one output device (e.g., a display device, printer or speaker). Such a system may also include one or more storage devices, such as disk drives, optical storage devices and solid-state storage devices such as random access memory (“RAM”) or read-only memory (“ROM”), as well as removable media devices, memory cards, flash cards, etc.
(293) Such devices also can include a computer-readable storage media reader, a communications device (e.g., a modem, a network card (wireless or wired), an infrared communication device, etc.) and working memory as described above. The computer-readable storage media reader can be connected with, or configured to receive, a computer-readable storage medium, representing remote, local, fixed and/or removable storage devices as well as storage media for temporarily and/or more permanently containing, storing, transmitting and retrieving computer-readable information. The system and various devices also typically will include a number of software applications, modules, services or other elements located within at least one working memory device, including an operating system and application programs, such as a client application or web browser. It should be appreciated that alternate embodiments may have numerous variations from that described above. For example, customized hardware might also be used and/or particular elements might be implemented in hardware, software (including portable software, such as applets) or both. Further, connection to other computing devices such as network input/output devices may be employed.
(294) Various embodiments may further include receiving, sending, or storing instructions and/or data implemented in accordance with the foregoing description upon a computer-readable medium. Storage media and computer readable media for containing code, or portions of code, can include any appropriate media known or used in the art, including storage media and communication media, such as, but not limited to, volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage and/or transmission of information such as computer readable instructions, data structures, program modules or other data, including RAM, ROM, Electrically Erasable Programmable Read-Only Memory (“EEPROM”), flash memory or other memory technology, Compact Disc Read-Only Memory (“CD-ROM”), digital versatile disk (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices or any other medium which can be used to store the desired information and which can be accessed by the system device. Based on the disclosure and teachings provided herein, a person of ordinary skill in the art will appreciate other ways and/or methods to implement the various embodiments.
(295) While the disclosed embodiments are described with respect to a single or biplane X-ray imaging modality, variations within these embodiments are also applicable for 3D reconstructions for instance based on rotational angiography, computed tomography, magnetic resonance imaging and the like. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense. It will, however, be evident that various modifications and changes may be made thereunto without departing from the broader spirit and scope of the invention as set forth in the claims.
(296) Other variations are within the spirit of the present disclosure. Thus, while the disclosed techniques are susceptible to various modifications and alternative constructions, certain illustrated embodiments thereof are shown in the drawings and have been described above in detail. It should be understood, however, that there is no intention to limit the invention to the specific form or forms disclosed, but on the contrary, the intention is to cover all modifications, alternative constructions and equivalents falling within the spirit and scope of the invention, as defined in the appended claims.
(297) The use of the terms “a” and “an” and “the” and similar referents in the context of describing the disclosed embodiments (especially in the context of the following claims) are to be construed to cover both the singular and the plural, unless otherwise indicated herein or clearly contradicted by context. The terms “comprising,” “having,” “including” and “containing” are to be construed as open-ended terms (i.e., meaning “including, but not limited to,”) unless otherwise noted. The term “connected,” when unmodified and referring to physical connections, is to be construed as partly or wholly contained within, attached to or joined together, even if there is something intervening. Recitation of ranges of values herein are merely intended to serve as a shorthand method of referring individually to each separate value falling within the range, unless otherwise indicated herein and each separate value is incorporated into the specification as if it were individually recited herein. The use of the term “set” (e.g., “a set of items”) or “subset” unless otherwise noted or contradicted by context, is to be construed as a nonempty collection comprising one or more members. Further, unless otherwise noted or contradicted by context, the term “subset” of a corresponding set does not necessarily denote a proper subset of the corresponding set, but the subset and the corresponding set may be equal.
(298) Operations of processes described herein can be performed in any suitable order unless otherwise indicated herein or otherwise clearly contradicted by context. Processes described herein (or variations and/or combinations thereof) may be performed under the control of one or more computer systems configured with executable instructions and may be implemented as code (e.g., executable instructions, one or more computer programs or one or more applications) executing collectively on one or more processors, by hardware or combinations thereof. The code may be stored on a computer-readable storage medium, for example, in the form of a computer program comprising a plurality of instructions executable by one or more processors. The computer-readable storage medium may be non-transitory.
(299) Preferred embodiments of this disclosure are described herein, including the best mode known to the inventors for carrying out the invention. Variations of those preferred embodiments may become apparent to those of ordinary skill in the art upon reading the foregoing description. The inventors expect skilled artisans to employ such variations as appropriate and the inventors intend for embodiments of the present disclosure to be practiced otherwise than as specifically described herein. Accordingly, the scope of the present disclosure includes all modifications and equivalents of the subject matter recited in the claims appended hereto as permitted by applicable law. Moreover, any combination of the above-described elements in all possible variations thereof is encompassed by the scope of the present disclosure unless otherwise indicated herein or otherwise clearly contradicted by context.
(300) All references, including publications, patent applications and patents, cited herein are hereby incorporated by reference to the same extent as if each reference were individually and specifically indicated to be incorporated by reference and were set forth in its entirety herein.