Method and device for processing ultrasound signal data

11331079 · 2022-05-17

Assignee

Inventors

Cpc classification

International classification

Abstract

The invention provides an ultrasound data processing method for pre-processing signal data in advance of generating ultrasound images. The method seeks to reduce noise through application of coherent persistence to a series of raw ultrasound signal representations representative of the same path or section through a body but at different successive times. A motion compensation procedure including amplitude peak registration and phase alignment is applied to raw echo signal data in advance of application of persistence in order to cohere the signals and thereby limit the introduction of motion induced artifacts.

Claims

1. A method of processing ultrasound signal data for use in generating ultrasound images, comprising: controlling, by a processor, an ultrasound transducer array to obtain raw echo signals, wherein the ultrasound transducer array is operatively coupled to the processor; receiving, by the processor, a plurality of raw echo signal representations of the raw echo signals, each of the plurality of raw echo signal representations representative of a same trajectory, section or volume through a body at different successive times; applying, by the processor, a motion compensation procedure to the plurality of raw echo signal representations, the motion compensation procedure comprising: designating, by the processor, a raw echo signal representation mid-way between a most recently received raw echo signal representation and a least recently received raw echo signal representation as a reference raw echo signal representation; registering, by the processor, amplitude peaks of remaining raw echo signal representations to amplitude peaks of the reference raw echo signal representation to thereby derive a plurality of registered raw echo signal representations; and aligning phases of the plurality of registered raw echo signal representations to one another to thereby derive a plurality of motion-compensated signal representations; applying an averaging technique to one or more subsets of the plurality of motion-compensated signal representations to thereby derive one or more averaged signal representations; generating, by the processor, one or more ultrasound images based on the one or more averaged signal representations; and outputting, by the processor, the one or more ultrasound images on a display operatively coupled to the processor.

2. The method as claimed in claim 1, wherein the motion compensation procedure comprises a two-step process, comprising first registering amplitude peaks of the plurality of raw echo signal representations to one another, and subsequently applying a phase correction procedure to align phases of the plurality of raw echo signal representations to one another.

3. The method as claimed in claim 1, wherein the one or more averaged signal representations comprise a plurality of averaged signal representations, and wherein the method further comprises: processing said plurality of averaged signal representations to generate a corresponding plurality of averaged envelope signal representations or ultrasound images; and applying a further averaging technique to said plurality of averaged envelope signal representations or ultrasound images.

4. The method as claimed in claim 1, wherein the plurality of raw echo signal representations are received non-simultaneously.

5. The method as claimed in claim 4, wherein, subsequent to deriving the one or more averaged signal representations, a motion compensation procedure is applied to said one or more averaged signal representations comprising at least registering amplitude peaks of the one or more averaged signal representations to those of a most recently received raw echo signal representation.

6. The method as claimed in claim 2, wherein the motion compensation procedure comprises: processing the plurality of raw echo signal representations to derive a corresponding plurality of envelope signal representations; determining deviations between the plurality of envelope signal representations; performing adjustments to the plurality of raw echo signal representations based on the determined deviations, so as to register the amplitude peaks of the plurality of raw echo signal representations with one another and to thereby derive the plurality of registered raw echo signal representations; and applying the phase correction procedure to the plurality of registered raw echo signal representations so as to register the phases of the plurality of registered raw echo signal representations to one another.

7. The method as claimed in claim 6, wherein said determining deviations between the plurality of envelope signal representations comprises: determining deviations between temporally consecutive pairs at least a subset of the plurality of envelope signal representations, said the subset including one envelope signal representation designated as a reference envelop signal representation; adding, to each of the deviations between temporally consecutive pairs, a sum of all preceding or succeeding determined deviations up to the reference envelope signal representation, such as to derive, for each envelope signal representation, cumulative deviations with respect to the reference envelope signal representation, and adjusting the plurality of raw echo signal representations, based on the determined cumulative deviations for the plurality of envelope signal representations, to thereby achieve registration of the amplitude peaks of the plurality of raw echo signal representations.

8. The method as claimed in claim 1, wherein the motion compensation procedure comprises: determining deviations between signal the amplitude peaks of the plurality of raw echo signal representations; performing adjustments to the plurality of raw echo signal data representations based on said determined deviations, so as to: achieve registration of the amplitude peaks of the different plurality of raw echo signal representations to one another and derive the plurality of registered raw echo signal representations, and alignment of the phases of the plurality of registered raw echo signal representations to one another.

9. The method as claimed in claim 2, wherein said phase correction procedure comprises: determining a phase shift between at least a subset of the plurality of raw echo signal representations and the reference raw echo signal representation; and adjusting the phases of each of the plurality of raw echo signal representations in accordance with the determined phase shifts.

10. The method as claimed in claim 1, further comprising: determining, following said motion compensation procedure, a measure of similarity between the plurality of raw echo signal representations; and determining, based on said measure of similarity, a total number of raw echo signal representations to be included in each of said one or more subsets to which the averaging technique is applied.

11. The method as claimed in claim 1, wherein said averaging technique comprises determining a weighted average of the plurality of motion compensated signal representations, and wherein the method further comprises: determining, following application of said motion compensation procedure, a measure of similarity between each raw echo signal representation and the reference raw echo signal representation; and determining a weighting factor for each raw echo signal representation based on the determined measure of similarity for said raw echo signal representation, the weighting factors for use in determining said weighted average of the plurality of motion compensated signal representations.

12. The method as claimed in claim 1, wherein each raw echo signal representation comprises data corresponding to a plurality of individual echo signals.

13. The method as claimed in claim 1, further comprising wherein generating the ultrasound image comprises processing the one or more averaged data signal representations to form the one or more ultrasound images.

14. An ultrasound diagnostic imaging system, comprising: an ultrasound transducer array; and a processor operatively coupled with the ultrasound transducer array, wherein the processor is adapted to: control the ultrasound transducer array to obtain raw echo signals; receive a plurality of raw echo signal representations of the raw echo signals, each of the plurality of raw echo signal representations representative of a same trajectory, section or volume through a body at different successive times; apply a motion compensation procedure to the plurality of raw echo signal representations, the motion compensation procedure comprising: designating a raw echo signal representation mid-way between a most recently received raw echo signal representation and a least recently received raw echo signal representation as a reference raw echo signal representation; registering amplitude peaks of remaining raw echo signal representations to amplitude peaks of the reference raw echo signal representation to thereby derive a plurality of registered raw echo signal representations; and aligning phases of the plurality of registered raw echo signal representations to one another to thereby derive a plurality of motion-compensated signal representations; apply an averaging technique to one or more subsets of the plurality of motion-compensated signal representations to thereby derive one or more averaged signal representations; generate one or more ultrasound images based on the one or more averaged signal representations; and output the one or more ultrasound images on a display operatively coupled to the processor.

15. The method as claimed in claim 9, wherein a phase shift for each given raw echo signal representation f.sub.i is determined based on the expression α=∠(K.Math.(f.sub.0.Math.f.sub.i*)), where α is a set of phase shifts, K is a convolution kernel, and f.sub.0 is the reference raw echo signal representation.

16. The method as claimed in claim 10, wherein the total number of the raw echo signal representations included in each of said one or more subsets to which the averaging technique is applied is positively related to said determined measure of similarity.

17. The method as claimed in claim 11, wherein said weighting factors are binary weighting factors, wherein each motion compensated signal representation is only included as part of the weighted average in the case that its measure of similarity with respect to the reference raw echo signal representation is above a certain threshold.

18. The method as claimed in claim 12, wherein said plurality of individual echo signals are representative of a set of adjacent paths through said body, so as to form a representation of a section or a volume through the body.

Description

BRIEF DESCRIPTION OF THE DRAWINGS

(1) Examples of the invention will now be described in detail with reference to the accompanying drawings, in which:

(2) FIG. 1 shows relative signal to noise ratio increase for coherent and incoherent persistence;

(3) FIG. 2 shows block diagrams illustrating example coherent persistence processes in the absence of motion compensation;

(4) FIG. 3 shows a flow diagram illustrating an example coherent persistence process in the absence of motion compensation;

(5) FIG. 4 shows an example method in accordance with an embodiment of the invention, wherein a combination of coherent and incoherent persistence is applied;

(6) FIGS. 5-7 schematically illustrate in flow diagram form an example motion compensation algorithm as incorporated within one or more embodiments of the present invention;

(7) FIG. 8 schematically illustrates an example signal processing method in accordance with an embodiment of the present invention;

(8) FIG. 9 schematically illustrates an example signal processing method in accordance with the invention in which a weighted average is calculated based on determining similarity metrics;

(9) FIG. 10 schematically depicts a variation on one or more embodiments of the present invention; and

(10) FIG. 11 schematically depicts an example ultrasound diagnostic imaging system.

DETAILED DESCRIPTION OF THE EMBODIMENTS

(11) The invention provides an ultrasound data processing method for pre-processing signal data in advance of generating ultrasound images. The method seeks to reduce noise through application of coherent persistence to a series of raw ultrasound signal representations representative of the same path or section through a body but at different successive times. A motion compensation procedure including amplitude peak registration and phase alignment is applied to raw echo signal data in advance of application of persistence in order to cohere the signals and thereby limit the introduction of motion induced artifacts.

(12) Embodiments of the invention may be implemented for instance within a signal processing unit of an ultrasound imaging system. Such a signal processing unit may preferably be configured to receive ultrasound echo signals directly and in real time from ultrasound transducers of an ultrasound probe (or from an associated signal pre-processing unit) and to process the received signal data for output to an image processing unit for instance. Alternatively, the method may be implemented ‘off-line’, in isolation from ultrasound imaging apparatus itself, for instance applied at a later time, after data capture has already occurred.

(13) As will be well-known to the skilled person, ultrasound signals may be applied to a body to be imaged, typically directed along a single ‘line’ or path through the body at any one time, for instance by a transducer array coupled to an ultrasound probe. The applied ultrasound stimuli are partially back-reflected at each boundary encountered as the signals pass through the body, leading to receipt at the transducer array of a series of echo back-reflections, received from different depths within the body. The time intervals between the different peaks in the echo-signals may be used to determine relative distanced between different medium boundaries in the body, from which a visual representation of the body may ultimately be constructed. Although the initially applied ultrasound pulse signal may be relatively short in duration, the resulting echo signal will typically be more temporally extended, as it will contain echo pulses received from a plurality of different depths within the imaged body and at different times depending upon the depth.

(14) As the skilled person will be aware, the echo signal received back at the transducer array in response to application of the ultrasound stimulus is commonly referred to in the art as the RF signal or RF data. RF signal is therefore a term of the art, and is not to be confused with the literal concept of radio frequency signals. The term may be used in the present disclosure to refer to the raw ultrasound signal data.

(15) Embodiments of the invention are based on application of persistence to the ultrasound echo signal data in a raw state; that is, in advance of any envelope-extraction. Such data may be the raw RF signal data, but may also be a complex IQ representation of the signal data and/or baseband IQ data. Any reference to RF data specifically should not be construed as limiting, and should be understood as replaceable by any other form of raw signal data (such as IQ data) without affecting the functioning of the described embodiment. The raw data may also be either fully sampled or down-sampled raw signal data.

(16) Embodiments of the invention are based on applying coherent averaging (coherent persistence) to RF echo signals corresponding to the same path or section through a body at a series of successive times. For brevity, in the present disclosure such RF signals of the same path/section at successive times will be referred to as ‘RF frames’. A single RF frame therefore refers to a data representation of one or more RF echo signals representative of a single path or section (i.e. plane) or volume through a body at a substantially singular time (or time interval). Where a frame is representative of a section or volume, the data representation may contain signal data corresponding to a plurality of individual adjacent paths forming said section.

(17) A series of echo signal representations corresponding to that same path or section at a plurality of consecutive time (intervals) may be referred to as a series of RF frames. The term ‘frame’ in this context is intended merely to capture the temporally consecutive nature of the echo signals, and should not be confused with the concept of an ‘image frame’ which refers to data processed such as to form a visual or graphical image.

(18) Coherent averaging or persistence in the present context means averaging or persistence applied to raw signal data or frames. Incoherent averaging or persistence in the present disclosure refers to averaging or persistence applied to envelope extracted data or ultrasound images. Furthermore, in the present disclosure, the terms image-based or envelope-based (incoherent) persistence are used to refer to persistence operated on both envelope data and persistence operated on data after logarithmic compression. The term RF-based persistence may be used interchangeably for coherent persistence, whether the persistence really uses RF data, IQ data or baseband IQ data. Results described in the present disclosure were primarily obtained from baseband IQ data at 5 MHz sampling.

(19) As discussed above, coherent persistence (that applied to raw signal data) achieves better signal-to-noise (SNR) improvement than incoherent persistence (that applied to envelope data or images).

(20) To quantitatively compare the difference between the SNR gain of coherent and incoherent persistence or averaging, a simple signal model will briefly be discussed enabling calculation of SNR. Let s.sub.k=S+.sub.k denote a (noise-corrupted) RF signal of RF frame k, where S is a constant representing the true signal, and n.sub.k complex zero-mean Gaussian noise of unity power. Without loss of generality, S is assumed real. The single-frame power SNR is therefore S.sup.2. The power SNR after averaging K RF frames is

(21) SNR RF , K = def signal power total power - signal power = S 2 E { .Math. 1 K .Math. K = 1 K ( S + n k ) .Math. 2 } - S 2 = S 2 [ S 2 + 1 K 2 E { .Math. .Math. K = 1 K n k .Math. 2 } ] - S 2 = S 2 [ S 2 + 1 K 2 K ] - S 2 = KS 2
where E is the expectation operator.

(22) Note that in the above denominator, since n.sub.k is a random noise of unity power, its sum over K frames adds to √{square root over (K)}. Hence E{|Σ.sub.K=1.sup.Kn.sub.k|.sup.2}=|√{square root over (K)}|.sup.2=K.

(23) In this case, the SNR increase from averaging is K. Converting to decibels using the well-known formula SNR=10 log.sub.10(SNR) dB, the SNR gain is equal to 10 log.sub.10K dB.

(24) The power SNR in the case of averaging over K envelope frames (i.e. incoherent averaging) is given by

(25) SNR envelope , K = def signal power total power - signal power = S 2 E { ( 1 K .Math. K = 1 K .Math. S + n k .Math. ) 2 } - S 2
FIGS. 1(a) and (c) show plots comparing numerical results for coherent and incoherent persistence in terms of resulting SNR (y-axis; dB) per number of frames averaged (x-axis), where each single frame has an SNR of 3 dB in FIG. 1(a) and 20 dB in FIG. 1(c). The upper curve in each plot corresponds to the SNR change for coherent (RF-frame) persistence, while the lower curve shows the SNR change for incoherent (envelope-frame) persistence. FIG. 1(b) and FIG. 1(d) show the difference in the resulting SNR (y-axis; dB) between coherent and incoherent averaging per number of frames averaged (x-axis) for each of the plots of FIG. 1(a) and FIG. 1(c) respectively.

(26) The curves show superior SNR improvement performance of RF-based persistence compared to envelope-based persistence. One source of this difference may be the fact that averaging over envelope frames cannot reduce noise-introduced signal-level-dependent bias in amplitude and therefore it is not effective in revealing the true signal amplitude, as opposed to RF-based averaging.

(27) The above model and sample results show that coherent (raw or RF signal based) persistence or averaging gives superior improvement in SNR compared to incoherent (envelope-signal) based persistence.

(28) The above-demonstrated improvement of SNR by coherent summation of data assumes that the collected signal representations maintain spatial coherence between the RF-frames to be averaged. However, when imaging tissue in vivo rather than a static phantom, relative motion between the ultrasound transducer and the body being imaged, if not managed properly, can degrade spatial coherence and result in deterioration in performance. The impact of motion increases in relation to the number of frames averaged in the persistence operation. This means that for coherent averaging, typically only very small averaging kernels can be used in order to avoid significant breakdown in spatial coherence between averaged RF frames and therefore deterioration in SNR improvement.

(29) However, to circumvent this, and enable larger averaging kernels, in accordance with embodiments of the present invention, the raw RF signal frames are first pre-processed with a motion compensation procedure in advance of application of persistence in order to mitigate any spatial incoherence introduced by relative motion of the body being imaged. This ensures that, upon averaging over multiple frames as part of the coherent persistence, motion artifacts introduced by destructive interference (arising by averaging non-aligned RF signals) is avoided.

(30) Optionally, a combination of motion-compensated RF persistence and subsequently applied standard incoherent persistence may be used.

(31) Simple example implementation of coherent persistence will now be described, before moving on to explain application of this within embodiments of the present invention.

(32) For coherent persistence, when averaging over K frames, the incremental gain of SNR for each extra frame averaged is most pronounced when K is small. This is evident in FIGS. 1(a) and (c) which show a logarithmic-like relationship between cumulative SNR and number of frames averaged. This suggests that the most pronounced gains in SNR from coherent persistence may be achieved at the very start of the persistence process. For this reason, it is desirable to apply coherent persistence at least over an initial few RF frames, even if not over all frames, so that a few dB of SNR gain can be easily achieved.

(33) One of the simplest forms of RF-based persistence is direct low-pass filtering, for example applied to beam-summed (i.e. beamformed) RF signal data over multiple ‘frames’, i.e. filtering a series of signals representative of different times or different time periods (or filtering ‘in slow time’ as it is often referred to in the art).

(34) The low pass filter may for instance be a simple as a moving average filter. An example process incorporating such persistence is shown in FIG. 2(a) wherein K RF frames R.sub.k are passed through a low pass filter in step 22, thereby producing filtered RF frames. The filtered RF frames are then post processed in step 24, for instance through envelope-extraction and/or image generation.

(35) In more complex examples, the simple coherent persistence 22 may be followed by a subsequent step of image-based persistence. FIG. 2(b) shows an example, wherein K RF frames R.sub.k are first passed through a moving average filter in step 22 to produce K filtered RF frames. The frames are then processed in step 26 with frequency compounding.

(36) Frequency compounding is a well-known procedure in the art. In brief, frequency compounding comprises the following steps. The raw RF data is independently filtered M times, each with a different bandpass filter. The bandpass filters differ in bandwidth and central frequency. These M separate filtration processes produce a set of M envelope datasets or images, each differing slightly, in particular in terms of their speckle since this is frequency dependent. Those M envelope datasets or images are then combined (i.e. averaged) incoherently.

(37) Since different RF frequencies within the raw data are expected to result in different realizations of the speckle, the process of incoherently averaging the M differently filtered datasets has the effect of averaging the speckle, thus lowering speckle variance and increasing image quality.

(38) Frequency compounding is not essential to the invention, and merely represents one example of a procedure that may be applied to the K filtered RF frames.

(39) Following this, the data is processed to form one or more images, and image-based (incoherent) persistence applied in step 28. This is hence an example of a ‘hybrid’ approach in which both coherent and incoherent persistence are used.

(40) Although frequency compounding is shown in the example of FIG. 2(b), this is by way of mere illustration only and is not essential to any persistence process used in embodiments of the present invention.

(41) FIG. 3 shows in schematic flow-diagram form the procedure applied in FIG. 2(b). The procedure is applied to an initial input series of five RF signal frames R.sub.n. A persistence process is applied consisting of application of a moving-average calculation procedure (via a moving average filter). As shown, this results in generation from the five input RF frames of three averaged frames R.sub.n′, each formed from an average of three of the input frames. For instance R.sub.1, R.sub.2 and R.sub.3 are each multiplied by ⅓ and then summed to form R′.sub.1.

(42) Each of the three averaged RF frames is then processed with frequency compounding (abbreviated as FC in FIG. 3). Following this, the RF signal frames are processed to form either envelope frames or image frames I.sub.1, I.sub.2, I.sub.3 and then a persistence process applied to the envelope or image frames to achieve average envelope or image frames I′.sub.1 and I′.sub.2. In the example shown in FIG. 3, the image-based persistence process comprises simply application of a moving-average procedure, wherein each of averaged image frames I′.sub.1 and I′.sub.2 are formed of an average of two of the frequency compounded image frames I.sub.1, I.sub.2, I.sub.3.

(43) The above discussion has illustrated some simple examples of coherent persistence as may be included as part of embodiments of the invention.

(44) However, such methods applied on their own would rely on operators of ultrasound imaging probes keeping the probes extremely still so as to manually minimize motion artifacts. On their own, the above procedures might be suitable for improving SNR over very short series' of signal frames (so called ‘weak’ persistence), where the risk of movement between frames within the body being imaged is less. However, even here, in the presence of local motion of physiological features or structures within the body, it might even be hard to maintain coherence between directly consecutive frames.

(45) To allow for a larger RF persistence kernel (i.e. a larger series of averaged RF frames) and to enhance robustness against physiological motion, motion compensation is, in accordance with embodiments of the invention, incorporated to register, or align, RF frames with respect to each other before averaging.

(46) The number of RF frames to be included in each averaging kernel (i.e. to be included in each averaged signal) can be adaptively determined in accordance with some examples, by a confidence measure that is derived during motion compensation.

(47) One example of a method in accordance with an embodiment of the invention is illustrated in FIG. 4.

(48) A series of K RF frames R.sub.k are received and are first processed in step 32 in accordance with a motion estimation and compensation procedure in order to reduce motion-induced deviations or displacements between the respective frames of the series and thereby register the RF frames to one another. A phase correction procedure is also applied. Following this, a coherent persistence procedure 22 is applied to the resulting motion-compensated frames R.sub.k′. The persistence procedure may for instance simply comprise a moving average of the frames (in in the example of FIG. 3) or may comprise a weighted average. The weightings may be derived based on an estimated degree of alignment or spatial coherence of each frame to all of the other frames. This will be described in greater detail below.

(49) The coherent persistence or averaging procedure 22 results in a set of motion compensated, averaged RF frames R″.sub.k. These may then be processed in step 26 with frequency compounding to realize K equivalent envelope frames E.sub.k comprising envelope extracted ultrasound signal data. Following this, the envelope frames may be processed in step 28 with non-coherent persistence, wherein an averaging procedure is applied to the envelope frames, resulting in a set of averaged envelope frames E′.sub.k.

(50) Optionally, during motion tracking, an extra step 38 may be performed to derive a confidence measure regarding how well the registration between different RF frames has been performed in step 32. A candidate for this confidence measure may be the cross-correlation coefficient among registered RF frames R′.sub.k. Based on the confidence measure, the optimal RF persistence kernel size (i.e. the optimal number of frames to be averaged) can be determined. The registration results can also be used to decide on the kernel size for the envelope-based persistence operation executed subsequently. Where the confidence measure is low, meaning the RF frames are not well registered, the size of the averaging kernel may be selected to be small, so as to minimize the risk of motion-induced artifacts in the averaged frames R″.sub.k and/or E′.sub.k.

(51) Although in the example of FIG. 4, the motion estimation and compensation step 32 and the coherent RF persistence step 22 are shown as separate, in alternative examples, they may be coupled or integrated into one single operation or method step. In step 32, motion estimation and compensation is performed on the raw RF frames. Many suitable algorithms exist for estimating motion, including standard Doppler processing (1-D), speckle tracking between RF frames, cross-correlation, or a sum of absolute difference procedure (2-D or 3-D). Speckle tracking is a well-known method in the present field and the skilled person would be aware of means for applying it. Tracking of movement between frames may also be determined based on gradient descent or optical flow.

(52) Gradient descent and optical flow are iterative optimization techniques that aim to find the deformation or displacement field between RF frames that minimizes the difference of intensity values between a reference frame and a frame which is warped to the reference frame. See for instance O. Somophone, “Fast Myocardial Motion and Strain Estimation in 3D Cardiac Ultrasound”, 2013.

(53) Tracking motion with higher dimensional data is also an option. For example, in the case of 2-D imaging, better robustness against noise or interference can be achieved by using 2-D tracking with 3-D (2-D spatially plus 1-D temporally) data, especially at deeper imaging depths within the body where signal-to-noise ratio is typically low to begin with. Motion tracking with four dimensional data can also be performed (i.e. 3-D spatially, plus 1-D temporally).

(54) As an alternative to directly registering the RF frames to one another, one efficient approach is to instead apply a two-step method comprising first determining deviation or displacement between the corresponding envelope signals for the frames, and then subsequently applying these determined deviations to the original raw RF frames in order to warp the frames to one another and achieve registration. Displacement estimation between envelope signals is a well-known task in ultrasound imaging. One particularly fast and efficient algorithm for achieving this, in particular in the case of real-time signal analysis and imaging, is the “Spare Demons” tracking algorithm. This is described in detail for instance in O. Somphone, “Fast Myocardial Motion and Strain Estimation in 3D Cardiac Ultrasound”, 2013.

(55) The algorithm takes as input the corresponding envelope signal representations/frames E.sub.0, E.sub.1 of at least two RF signal frames R.sub.0, R.sub.1. These are derived for instance by taking a copy of the RF signal frames and then applying a procedure to extract the envelope data, from which corresponding envelope frames may be constructed. One of the envelope signal frames E.sub.0 is denoted as a ‘reference’ frame and the other E.sub.1 as a ‘template’ frame.

(56) An estimated displacement (or displacement ‘field’) is determined between the reference frame E.sub.0 and the template frame E.sub.1. This displacement corresponds to motion of the body being imaged between the successive frames.

(57) Once the displacement field between the reference envelope frame and template envelope frame is calculated, this may then be added or otherwise applied to corresponding original RF template frame so as to register the RF template frame to the RF reference frame. This process is known as ‘warping’ the template frame to the reference frame.

(58) Once the template frame has been registered or warped to the reference frame, the phase of the template frame must, for this embodiment, be compensated to align with that of the reference frame. A residual phase between the two frames is calculated, and the phase of the template frame adjusted accordingly.

(59) The process may be represented in equations as follows. Consider two RF signal frames R.sub.0, R.sub.1 to be registered to one another. First, the corresponding envelope-signal representations or frames E.sub.0, E.sub.1 are extracted (representing variation of amplitude of the RF signals).

(60) The task is to register R.sub.1 to R.sub.0.

(61) Displacement estimation is first performed between the envelope frames E.sub.0 and E.sub.1. The problem of displacement estimation is to find the motion field u(x) that minimizes
E{|E.sub.0(x)−E.sub.1(x+u(x))|.sup.2}
Any appropriate technique may be performed to find the displacement field u(x), including for instance optical flow, gradient descent, or any other technique. Note that so far only the envelopes have been used.
Based on the displacement field u(x), the RF frame R.sub.1 must be warped to the reference RF frame R.sub.0.
Warping the RF data is performed by interpolating the frame R.sub.1 as follows:
R.sub.1.sub.warp(x)=R.sub.1(x+u(x))
Note that this operation applies to the RF (i.e. raw) data.
Following this, residual phase between R.sub.0 and R.sub.1 is compensated. The residual phase α(x) may be determined as follows:
α(x)=angle(E{R.sub.0(x)R.sub.1.sub.warp(x)*}
The phase compensation of R.sub.1 to match R.sub.0 may then be performed as follows:

(62) R 1 warp_phase compensated ( x ) = R 1 warp ( x ) e i α ( x )
Now R.sub.0 and R.sub.1_warp_phase compensated can be summed (coherent summing) and a coherent average subsequently derived (coherent persistence). Their structures will align (based on the warping with u(x)) and their phases will be aligned (based on the phase compensation).
Examples of procedures for reducing deviation between an entire series of RF frames (i.e. registering a series of RF frames) will now be described in detail.

(63) Embodiments of the invention are based on the concept of applying persistence to a consecutive series of echo signal frames after application of motion compensation. In applying motion compensation to such a series of frames, it is desired that all frames should be aligned (or registered) to one another. One efficient approach to achieving this is to designate a single frame in the series as a reference frame and to determine a deviation between every other frame and this reference frame in respect of certain features or characteristics. Each frame is then warped or adjusted in accordance with its respective determined deviation from the reference frame such as to eliminate the deviation and render it better aligned or cohered with the reference frame. Warping each frame according to a deviation with the same single reference frame, rather than for instance with respect to an adjacent frame, has been found to be more accurate, since the deviation tends to be greater for more temporally disparate frames, and hence the margin of error in its determination less.

(64) Although in preferred examples a deviation between each frame and a reference frame is ultimately determined, this may, in accordance with one or more examples, be derived through an algorithm that is based upon determining deviations between consecutive frames, and then appropriately summing these in order to determine deviations between each frame and the reference frame.

(65) Examples of this approach are illustrated in FIGS. 5 and 6. In this example, as in the example described above, registration is performed by first determining deviations between the corresponding envelope-signal frames for a series of RF frames, and then adjusting the RF frames in accordance with these determined envelope frame deviations. FIGS. 5 and 6 illustrate the process of determining the deviations between the envelope-signal frames E.sub.n corresponding to the series of raw RF frames. The figures show the series of envelope frames E.sub.n, the inter-frame deviations D between consecutive frames, and the cumulative deviations CD between each of the frames and one of the frames designated as a reference frame.

(66) In accordance with a first approach, illustrated in FIG. 5, the reference frame may be designated as the temporally first envelope-signal frame in the series. In the case illustrated, this is envelope-signal frame E.sub.N−3. The deviations D.sub.n between each consecutive pair of envelope-signal frames is determined. Following this, these deviations are successively summed so as to provide a respective cumulative deviation CD.sub.n for each signal frame apart from the reference frame E.sub.N−3. These indicate a total deviation between each frame E.sub.n and the reference frame.

(67) As shown, the first cumulative deviation CD.sub.N−2 for frame E.sub.N−2 consists simply of the inter-frame deviation D.sub.N−2. The cumulative deviation CD.sub.N−1 for frame E.sub.N−1 is equal to the sum of inter-frame deviations D.sub.N−2 and D.sub.N−1. The cumulative deviation CD.sub.N for frame E.sub.N (the most temporally recent frame in the series) is equal to the sum of all three inter-frame deviations D.sub.N−2, D.sub.N−1, and D.sub.N.

(68) Following determining of the cumulative deviations between each of envelope-signal frames E.sub.N−2, E.sub.N−1, E.sub.N and the designated reference frame (temporally first envelope-signal frame E.sub.N−3), these cumulative deviations are used to warp each of the original raw RF signal frames R.sub.N−2, R.sub.N−1, R.sub.N (not shown) to the corresponding reference RF frame R.sub.N−3 so as to eliminate the deviations. Warping may be achieved in accordance with the procedure outlined above, wherein each RF frame is interpolated based on its respective cumulative deviation or displacement. In other words, R.sub.n warp (x)=R.sub.n(x+CD.sub.n(x)), where CD.sub.n(x) is the cumulative deviation field used to warp the RF frame and x is the grid of sample points (where this may be a time dimension, i.e. t, for a raw RF signal for instance).

(69) In accordance with a second example, schematically depicted in FIG. 6, the most temporally recent envelope-signal frame (in the case illustrated, E.sub.N) is designated as the reference frame. As in the example of FIG. 5, inter-frame deviations D.sub.N−1, D.sub.N−2 and D.sub.N−3 are determined between each consecutive pair of envelope-signal frames. Following this, a set of cumulative deviations CD.sub.N are calculated, this time starting from the most recent frame E.sub.N and working backwards to the first signal frame E.sub.N−3. Each of the cumulative deviations CD.sub.N thereby gives a total deviation between each of the frames E.sub.N−1, E.sub.N−2 and E.sub.N−3 and the reference frame E.sub.N.

(70) Again, once the cumulative deviations are calculated, each of the corresponding original raw RF frames R.sub.N−1, R.sub.N−2 and R.sub.N−3 may be warped to the designated reference RF frame R.sub.N based on the envelope deviations in the manner described above.

(71) In accordance with either of the above approaches (of FIG. 5 or FIG. 6), calculation of the inter-frame deviations D.sub.N+1 and/or the cumulative deviations CD.sub.n may be performed all at once, for instance after collection of all frames which are to be included in the series. Alternatively, the inter-frame deviations and cumulative deviations may be determined ‘in real time’, such that these are appended and/or updated each time a new signal frame is added to the series.

(72) In accordance with the example of FIG. 5, this would require, upon appending of each new frame RF frame R.sub.n, determination of the corresponding envelope frame E.sub.n and then calculation of the intra-frame deviation between said envelope frame and the immediately preceding envelope frame E.sub.n−1. It would then require calculation of a cumulative deviation for the appended envelope frame with respect to the reference envelope frame (in this case E.sub.N−3). This would simply require adding the newly determined inter-frame deviation D.sub.n+1 to the previous three inter-frame deviations, or equivalently, to the immediately preceding cumulative deviation CD.sub.n. This process would be repeated for each new frame appended to the series. The newly updated deviations may then be reapplied to the corresponding series of original raw RF frames to achieve motion compensation.

(73) In accordance with the example of FIG. 6, ‘real time’ calculation of the deviations would be more complex, requiring updating of all previously calculated cumulative deviations. This is because in the example of FIG. 6, the reference frame is designated as the most recent frame. Hence, if a new frame is added, the reference frame has changed and previously calculated cumulative deviations are no longer accurate. However, the example of FIG. 6 may be preferred, despite the possible slight additional computational cost, since it reduces apparent ‘time lag’ in subsequently generated images. Since all RF frames are warped to the temporally most recent frame, the resulting average provides a representation of the imaged body at the time of the most recently captured signal, rather than a time, say two or three frames in the past.

(74) For the example of FIG. 6, upon appending of a newly acquired RF frame R.sub.N+1, new envelope frame E.sub.N+1 is extracted. The inter-frame deviation D.sub.N between this frame and the immediately preceding frame E.sub.N is then calculated. Following this, each of the previously calculated cumulative deviations CD.sub.N−3, CD.sub.N−2, and CD.sub.N−1 must be updated by adding to them the newly calculated inter-frame deviation D.sub.N+1. This in effect updates all of the cumulative deviations such that they each provide the total deviation between each of the respective envelope frames and the newly added envelope frame E.sub.N+1. Finally, a new cumulative deviation must be added for frame E.sub.N (which is now no longer the reference frame), this being equal simply to the new inter-frame deviation D.sub.N between new frame E.sub.N+1 and immediately preceding frame E.sub.N. The resulting, updated, frame structure is shown in FIG. 7.

(75) The above described process achieves efficient registering of RF frames to one another so as to align amplitude peaks and reduce the risk of motion artifacts. However, as discussed above, even when the RF frames have been warped or adjusted such that their respective envelopes' features are matched, the RF signals comprised by said frames are not necessarily coherent with one another. Summing them as such when motion has occurred generally results in black stripes and cracks due to destructive interferences.

(76) To eliminate these artifacts, a phase adjustment procedure must also be performed.

(77) In the phase adjustment procedure, the local phase shift at every sample (of each RF signal) of each warped RF frame f.sub.i with respect to the corresponding samples of the designated reference frame f.sub.0 is first estimated. The phase shift is then corrected for instance through use of a suitable convolution kernel K. This similar for instance to a Doppler estimation and can be performed for instance on down-sampled IQ data.

(78) To estimate the local phase shift for every RF signal of each RF frame, the following estimation expression may be used:
α=∠(K.Math.(f.sub.0.Math.f.sub.i*))  (1)
where f.sub.0 and f.sub.i are complex representations of the reference and warped RF frames respectively.

(79) Adding the quantity derived from (1) to the warped template's phase ensures constructive interference with the reference, and therefore mitigation or elimination of motion artifacts arising due to phase incoherence:
f.sub.i′=f.sub.i.Math.e.sup.iα  (2)

(80) Since this procedure is aimed at remedying spatial displacements between two lines or planes of focus of two respective RF frames using phase adjustment, the procedure only makes sense where the spatial displacement is less than the length of a single pulse wavelength, and ideally under a quarter of a pulse wavelength (half a wavelength in round trip) in the axial direction and under one (A-line) path width in lateral dimension.

(81) Although the warped template and the reference do not necessarily always fulfill these conditions, the resulting average, when the RF data is rendered into an image, still looks relatively sharp and defect free up to a certain inter-frame distance.

(82) The frame warping (amplitude peak adjustment) and phase adjustment may be thought of as two parts of an alignment procedure, operating at different scales, or precision levels. Frame warping aligns the envelope peaks, but leaves errors in the signal alignment on the order of ¼ wavelength. Phase adjustment remedies these residual small-scale misalignments.

(83) FIG. 8 schematically outlines the full procedure for processing a series of RF frames in accordance with an embodiment of the invention. As discussed above, the term RF frame in the present disclosure is used by way of shorthand to denote a data representation of one or more raw (RF) echo signals representative of a given path or section through a body being imaged at a particular time. The series of RF frames represent said same path or section through the body at a series of respective time points. Each RF frame may comprise data representative of multiple adjacent paths (or multiple ‘lines’) forming a section through the body.

(84) In the example processing method of FIG. 8, a series of four RF frames, R.sub.n (shown at (b)), are applied first with a motion compensation procedure (stages (a)-(d)), then a phase correction procedure (stage (e)) and then finally a summing and averaging procedure (stage (f)).

(85) In accordance with a first stage (a), each RF frame R.sub.n is processed to extract a corresponding envelope signal E.sub.n. As in the motion tracking procedure described above with reference to FIGS. 5-7, once the envelope data has been extracted, consecutive, inter-frame deviations D.sub.n between each neighboring pair of envelope frames E.sub.n is then calculated (in step 32a).

(86) One of the RF frames R.sub.n (and corresponding envelope frame E.sub.n) is designated as a reference frame. In this case, the reference frame is designated to be R.sub.N (and E.sub.N) (the top frame shown in FIG. 8). Once the consecutive, inter-frame deviations have been determined, cumulative deviation CD.sub.n are calculated (in stage (c)) giving the total deviation between each envelope frame E.sub.n and the reference envelope frame E.sub.N. These are generated (as shown in step (c)) by successively summing the inter-frame consecutive deviations D.sub.n previously calculated, starting from D.sub.N−1 and continuing downwards to D.sub.N−3. The result of this is a set of three cumulative deviations, CD.sub.N−1, CD.sub.N−2, and CD.sub.N−3 corresponding respectively to the total deviation between the each of envelope frames E.sub.N−1, E.sub.N−2, and E.sub.N−3 and the reference envelope frame E.sub.N.

(87) Following calculation of the cumulative deviations CD.sub.n, in stage (d), each of the original, raw RF frames R.sub.n (excluding the reference frame R.sub.N) is warped or adjusted to the reference frame, based on the calculated cumulative deviations between the envelope frames E.sub.N. This warping (or adjusting) procedure is shown as step 32b, and generates as output a set of warped RF frames R′.sub.n corresponding to each of RF frames R.sub.N−1, R.sub.N−1, R.sub.N−1 warped to the reference frame R.sub.N to substantially eliminate the calculated deviations CD.sub.n.

(88) Following generation of the warped RF frames R′.sub.n, in stage (e), a phase correction procedure 36 is then applied (for example, as outlined in more detail above) so as to adjust the phases of each of the warped RF frames to align with that of the reference frame R.sub.N. The output of the phase correction procedure 36 is a set of three fully motion-compensated RF frames R″.sub.n.

(89) Finally, in stage (f), the motion compensated RF frames R″.sub.n are each added to the reference frame, to form a coherent sum J.sub.n of reference frames. A coherent average may be derived simply by applying a relevant weighting factor to each of the motion compensated frames R″.sub.n before summing.

(90) Preferably, the above process is done in a stepwise manner, wherein each and every stage ((a)-(f)) of the process is performed for a given frame, before moving on to the next frame and repeating the process Advantageously, this allows new frames to be added to the series, and for the process to be easily updated to accommodate the new frame. With each frame for which the process (stages (a)-(f)) is performed, the resulting motion compensated frame R″.sub.n is added to a running partial sum J.sub.n, representing a sum of all compensated frames so far calculated (or indeed, so far received).

(91) Where this step-wise approach is followed, preferably, in the phase correction stage (e), the phase adjustment of each warped frame R′.sub.n is performed ‘indirectly’, with respect to the phase of the partial sum J.sub.n, rather than with respect to the reference frame R.sub.N. This ensures that coherence is maintained at all stages of the procedure. However, direct phase adjustment with respect to the reference frame may alternatively be performed.

(92) It is also possible, similarly, to perform the frame warping of each successive RF frame R.sub.n to the coherent running average J.sub.n, rather than to a particular reference frame. In this case cumulative deviations are calculated between each frame and the latest partial sum J.sub.n.

(93) The advantage in the above procedure of FIG. 8 of calculating first the inter-frame deviations D.sub.n and then the cumulative deviations CD.sub.n (between each frame and the reference frame) is that the resulting algorithm is more efficient for application to real-time data processing where additional new RF frames are repeatedly appended to the series. By way of example, for each new frame R.sub.n+1 added to the series, the following steps may be performed, to update each of the motion and phase compensated frames and update the averaging sum J.sub.n.

(94) New RF frame R.sub.N+1 is received.

(95) This frame is designated as the new reference frame.

(96) This frame is processed to extract the envelope-signal to form corresponding envelope signal frame E.sub.N+1.

(97) The inter-frame displacement D.sub.N is calculated between the new envelope frame E.sub.N+1 and envelope frame E.sub.N.

(98) This new inter-frame displacement D.sub.N is added to every existing cumulative displacement CD.sub.n (that is, to each of CD.sub.N−1, CD.sub.N−2, CD.sub.N−3). The result is a set of new cumulative displacements giving the displacement between each frame and new frame R.sub.N+1. Each of frames R.sub.N to R.sub.N−3 is warped 32b to the new reference frame R.sub.N+1 based on the updated cumulative deviations.

(99) The warped frames are each re-processed 36 to correct any phase difference with respect to the newly added frame.

(100) The coherent sum J.sub.n is recalculated by summing each of the new warped, phase-adjusted frames

(101) Angle correction (phase adjustment) 36 is required regardless of the number of frames included in the coherent sum or average J.sub.n, in order that motion artifacts be effectively reduced. As noted above, to maintain coherence at every stage of the summation (stage (f)), preferably the angle correction 36 is applied with respect to the phase of the partial summation J.sub.n, rather than directly with respect to the reference frame. However, this is not essential, and the alternative is also possible.

(102) Assuming approximately 2 cm/s lateral velocity (of the body being imaged), and a frame rate of 85 captured frames per second, the above method is able to achieve up to 5 frames of persistence (i.e. an averaging Kernel of length five frames). This corresponds to a theoretical signal-to-noise ratio (SNR) increase of approximately 7 dB.

(103) In accordance with one or more embodiments, a hybrid coherent-incoherent persistence method may be performed, wherein the above procedure of FIG. 8 is performed upon each of a plurality of subseries of RF frames, to thereby generate a plurality of coherently averaged RF frames. This plurality of averaged frames may then be processed to extract corresponding envelope frames, and these envelopes then (incoherently) averaged to thereby derive a final hybrid coherent-incoherent averaged signal frame.

(104) In accordance with one or more embodiments of the presently claimed data processing method, the length of the averaging kernel for the persistence process (i.e. the number of RF frames included in the average) may be dynamically adapted in accordance with one or more factors to improve results (as outlined below).

(105) According to an existing approach for partially countering the effects of motion in persistence (where true motion compensation is not applied), the impact of the motion upon the obtained signal data is first assessed (for instance by checking cross-correlation values between the frames), and the length of the persistence (averaging) kernel is then shortened accordingly. For example, where it is determined that correlation between signal frames is low, the averaging kernel may be reduced in size so as to mitigate the risk of introducing motion artifacts. This helps to ensure artifact-free images, while still achieving relatively high SNR for steady observations. (Of course, this comes at the cost of reducing the effectiveness of the persistence process at reducing noise for non-steady observations).

(106) This approach may be advantageously incorporated into embodiments of the present invention (in which true motion compensation is performed) so as to reduce the risk of motion artifacts caused by still imperfect motion compensation. Here, the similarly between each warped frame (termed a ‘template frame’ for the following discussion) and the reference frame can be measured, and the warped frame may be added to the coherent average only if the similarity is great enough.

(107) Alternatively, the measured similarity may be used to determine a weighting factor for applying the template frame within the final average. For example, a poorly correlated warped template would be given a low weighting. Weightings may similarly be determined and applied to every frame added to the average, so as to realize a resulting average in which motion artifacts are minimized.

(108) An example of such a process is illustrated schematically in block diagram form in FIG. 9. Here, motion tracking and compensation 32 with respect to a reference frame R.sub.0 is first applied to template frame to realize warped frame R′.sub.1, and subsequently, phase adjustment 36 is performed with respect to reference frame R.sub.0 to realize fully motion compensated frame R″.sub.1. A similarity operation {circumflex over (X)} is then applied to determine a similarity measure σ between R.sub.0 and R′.sub.1. Based on this similarity measure, a weighting W is derived for multiplication with frame R″.sub.1 in order thereby to derive weighted frame R.sup.(3).sub.1. The weighted frame is then added to the final sum. By applying this process to each frame to be added to the sum, good results in terms of reducing motion artifacts may be achieved since strongly non-correlated frames are dampened in terms of their effect within the average.

(109) It is noted that a simple (motion-compensated) average consisting of just two consecutive frames is in most cases unlikely to lead to any significant motion artifacts, even where cross-correlation is relatively low. Hence, for improved computational efficiency, the similarity between the first two frames in any series to be averaged may, in accordance with examples, be ignored. Ignoring this metric only for the second frame should not degrade the resulting images. By contrast, applying a low weighting value to the second frame (where the similarity metric was low) would in fact cause significant drops in the achieved SNR improvement.

(110) In addition to motion, signal noise may also contribute to poor frame correlation. Noise levels generally increase with depth of tissue imaged. Hence, the cross-correlation matrix applied for realizing similarity between two frames may, in accordance with one or more examples, be normalized in accordance with imaging depth in advance of computing the weights W.

(111) In accordance with examples, where a hybrid mix of coherent and incoherent averaging is used (as discussed above), the cross-correlation metric between motion-compensated RF frames may be calculated and used to inform a determination as to how many frames should be coherently and incoherently averaged respectively. Also, different metrics (e.g. cross-correlation of RF data vs. cross-correlation of envelope data) could be used separately to determine the respective lengths of each of the coherent and incoherent averaging kernels.

(112) As discussed above, motion compensation in accordance with embodiments of the present invention may comprise warping each of the RF frames in the series to one designated reference frame. Typically, this frame is designated as the most recently captured frame or most recently received frame. This is illustrated schematically in FIG. 10 which shows that all frames extending back to R.sub.n−6 are warped to the latest captured frame R.sub.n. This is typically preferred since, in the case of real-time data processing, warping to less recent frames causes an apparent delay or time lag in observable ultrasound images (if these are being generated from the warped RF frames in real-time). For instance, deliberate movement of the probe or changes in the scene being observed are reflected in the generated images only a few seconds later.

(113) However motion tracking is typically less accurate when performed between frames which are greatly temporally separated. In the zero-latency case (i.e. the case where the most recent frame is the reference to which all others are warped), the maximum leap distance is equal to the entire size of the averaging kernel.

(114) In accordance with an alternative configuration therefore, the reference frame may instead be designated as a frame mid-way between the most recent frame in the series and the least recent frame in the series. This is illustrated in FIG. 10. Here, the reference frame R.sub.n is selected as a frame mid-way between the least recent frame R.sub.n−3 and the most recent frame R.sub.n+3.

(115) In exchange for a few frames of latency (a few frames of time delay in rendered images), warping towards the center instead of the end reduces the temporal separation between frame R.sub.n−6 and R.sub.n by half. The quality of the motion compensation is thereby increased. It should be noted that the computational cost of this configuration is no greater than that of a configuration in which the most recent frame is chosen as the reference frame.

(116) For this implementation, assuming 2 cm/s lateral motion at 85 frames per second, very good performance is achieved both in terms of SNR increase and reduction of motion artefacts for 9 frames of persistence. However, as a trade-off, these settings do generate approximately 50 ms of latency in any generated ultrasound images.

(117) As a further improvement upon the above method, a final step may be added to the process wherein the resulting coherent sum of the RF frames warped to the center or middle frame is then itself finally warped to the most recent frame. In this example, the steps are as follows. Each RF frame, as in the case above, is warped to the center frame in the series. The warped frames are then summed to generate a coherent sum of the warped RF frames. This coherent sum is then itself warped to the most recent frame (i.e. R.sub.n+3) in the example of FIG. 10. This will fully eliminate the latency while keeping the benefits of the symmetrical kernel in terms of robustness to motion artifacts.

(118) In all embodiments of the invention described, the RF frames (or data representations) may be representative of RF signal data which has been down-sampled or decimated. This process can introduce additional phase de-coherence, and phase information is lost when the down-sampling is performed. Image warping is generally based upon interpolation and resampling of data. Where phase information has been lost, it can be very difficult to recover full coherency between any warped RF frame and the reference RF frame.

(119) In accordance with one or more embodiments of the invention, fully sampled data may be used for each RF frame, rather than for instance down-sampled data. Experimental evidence has shown that when using fully sampled data, fewer artifacts are generated when warping frames which have relatively large time lags between them. This means that, in embodiments described above in which RF frames are each warped to the most recent frame in the series (to minimize time lag), this procedure may be performed while producing fewer defects in resulting images.

(120) In accordance with an aspect of the invention, there is provided a diagnostic imaging system comprising a processing unit adapted to carry out a signal processing method in accordance with any embodiment of the invention.

(121) The general operation of an exemplary ultrasound diagnostic imaging system will now be described, with reference to FIG. 11, and with emphasis on the signal processing function of the system since this invention relates to the processing of the signals measured by the transducer array.

(122) The system comprises an array transducer probe 40 which has a CMUT transducer array 100 for transmitting ultrasound waves and receiving echo information. The transducer array 100 may alternatively comprise piezoelectric transducers formed of materials such as PZT or PVDF. The transducer array 100 is a two-dimensional array of transducers 110 capable of scanning in a 2D plane or in three dimensions for 3D imaging. In another example, the transducer array may be a 1D array.

(123) The transducer array 100 is coupled to a microbeamformer 42 in the probe which controls reception of signals by the CMUT array cells or piezoelectric elements. Microbeamformers are capable of at least partial beamforming of the signals received by sub-arrays (or “groups” or “patches”) of transducers as described in U.S. Pat. No. 5,997,479 (Savord et al.), U.S. Pat. No. 6,013,032 (Savord), and U.S. Pat. No. 6,623,432 (Powers et al.).

(124) Note that the microbeamformer is entirely optional. The examples below assume no analog beamforming.

(125) The microbeamformer 42 is coupled by the probe cable to a transmit/receive (T/R) switch 46 which switches between transmission and reception and protects the main beamformer 50 from high energy transmit signals when a microbeamformer is not used and the transducer array is operated directly by the main system beamformer. The transmission of ultrasound beams from the transducer array 40 is directed by a transducer controller 48 coupled to the microbeamformer by the T/R switch 46 and a main transmission beamformer (not shown), which receives input from the user's operation of the user interface or control panel 68.

(126) One of the functions controlled by the transducer controller 48 is the direction in which beams are steered and focused. Beams may be steered straight ahead from (orthogonal to) the transducer array, or at different angles for a wider field of view. The transducer controller 48 can be coupled to control a DC bias control 75 for the CMUT array. The DC bias control 75 sets DC bias voltage(s) that are applied to the CMUT cells.

(127) In the reception channel, partially beamformed signals are produced by the microbeamformer 42 and are coupled to a main receive beamformer 50 where the partially beamformed signals from individual patches of transducers are combined into a fully beamformed signal. For example, the main beamformer 50 may have 128 channels, each of which receives a partially beamformed signal from a patch of dozens or hundreds of CMUT transducer cells or piezoelectric elements. In this way the signals received by thousands of transducers of a transducer array can contribute efficiently to a single beamformed signal.

(128) The beamformed reception signals are coupled to a signal processing unit 52. This may be a signal processing unit in accordance with an aspect of the present invention. The signal processing unit 52 can process the received echo signals in various ways, such as band-pass filtering, decimation, I and Q component separation, and harmonic signal separation which acts to separate linear and nonlinear signals so as to enable the identification of nonlinear (higher harmonics of the fundamental frequency) echo signals returned from tissue and micro-bubbles. The signal processing unit may also perform additional signal enhancement such as speckle reduction, signal compounding, and noise elimination. The band-pass filter in the signal processor can be a tracking filter, with its pass band sliding from a higher frequency band to a lower frequency band as echo signals are received from increasing depths, thereby rejecting the noise at higher frequencies from greater depths where these frequencies are devoid of anatomical information.

(129) The signal processing unit 52 is also adapted to perform one or more embodiments of the processing method of the present invention.

(130) The beamformers for transmission and for reception are implemented in different hardware and can have different functions. Of course, the receiver beamformer is designed to take into account the characteristics of the transmission beamformer. In FIG. 11 only the receiver beamformers 42, 50 are shown, for simplicity. In the complete system, there will also be a transmission chain with a transmission micro beamformer, and a main transmission beamformer.

(131) The function of the micro beamformer 42 is to provide an initial combination of signals in order to decrease the number of analog signal paths. This is typically performed in the analog domain.

(132) The final beamforming is done in the main beamformer 50 and is typically after digitization.

(133) The transmission and reception channels use the same transducer array 40′ which has a fixed frequency band. However, the bandwidth that the transmission pulses occupy can vary depending on the transmission beamforming that has been used. The reception channel can capture the whole transducer bandwidth (which is the classic approach) or by using bandpass processing it can extract only the bandwidth that contains the useful information (e.g. the harmonics of the main harmonic).

(134) The processed signals are coupled to a B mode (i.e. brightness mode, or 2D imaging mode) processing unit 56 and a Doppler processor 58. The B mode processor 56 employs detection of an amplitude of the received ultrasound signal for the imaging of structures in the body such as the tissue of organs and vessels in the body. B mode images of structure of the body may be formed in either the harmonic image mode or the fundamental image mode or a combination of both as described in U.S. Pat. No. 6,283,919 (Roundhill et al.) and U.S. Pat. No. 6,458,083 (Jago et al.) The Doppler processor 58 processes temporally distinct signals from tissue movement and blood flow for the detection of the motion of substances such as the flow of blood cells in the image field. The Doppler processor 58 typically includes a wall filter with parameters which may be set to pass and/or reject echoes returned from selected types of materials in the body.

(135) The structural and motion signals produced by the B mode and Doppler processors are coupled to a scan converter 62 and a multi-planar reformatter 74. The scan converter 62 arranges the echo signals in the spatial relationship from which they were received in a desired image format. For instance, the scan converter may arrange the echo signal into a two dimensional (2D) sector-shaped format, or a pyramidal three dimensional (3D) image. The scan converter can overlay a B mode structural image with colors corresponding to motion at points in the image field with their Doppler-estimated velocities to produce a color Doppler image which depicts the motion of tissue and blood flow in the image field. The multi-planar reformatter will convert echoes which are received from points in a common plane in a volumetric region of the body into an ultrasound image of that plane, as described in U.S. Pat. No. 6,443,896 (Detmer). A volume renderer 72 converts the echo signals of a 3D data set into a projected 3D image as viewed from a given reference point as described in U.S. Pat. No. 6,530,885 (Entrekin et al.).

(136) The 2D or 3D images are coupled from the scan converter 62, multi-planar reformatter 74, and volume renderer 72 to an image processor 60 for further enhancement, buffering and temporary storage for display on an image display 70. In addition to being used for imaging, the blood flow values produced by the Doppler processor 58 and tissue structure information produced by the B mode processor 56 are coupled to a quantification processor 64. The quantification processor produces measures of different flow conditions such as the volume rate of blood flow as well as structural measurements such as the sizes of organs and gestational age. The quantification processor may receive input from the user control panel 68, such as the point in the anatomy of an image where a measurement is to be made. Output data from the quantification processor is coupled to a graphics processor 66 for the reproduction of measurement graphics and values with the image on the display 70, and for audio output from the display device 70. The graphics processor 66 can also generate graphic overlays for display with the ultrasound images. These graphic overlays can contain standard identifying information such as patient name, date and time of the image, imaging parameters, and the like. For these purposes the graphics processor receives input from the user interface 68, such as patient name. The user interface is also coupled to the transmit controller 48 to control the generation of ultrasound signals from the transducer array 40′ and hence the images produced by the transducer array and the ultrasound system. The transmit control function of the controller 48 is only one of the functions performed. The controller 48 also takes account of the mode of operation (given by the user) and the corresponding required transmitter configuration and band-pass configuration in the receiver analog to digital converter. The controller 48 can be a state machine with fixed states.

(137) The user interface is also coupled to the multi-planar reformatter 74 for selection and control of the planes of multiple multi-planar reformatted (MPR) images which may be used to perform quantified measures in the image field of the MPR images. As discussed above, embodiments make use of a controller and processing unit. These can each be implemented in numerous ways, with software and/or hardware, to perform the various functions required. A processor is one example of a controller or processing unit which employs one or more microprocessors that may be programmed using software (e.g., microcode) to perform the required functions. A controller or processing unit may however be implemented with or without employing a processor, and also may be implemented as a combination of dedicated hardware to perform some functions and a processor (e.g., one or more programmed microprocessors and associated circuitry) to perform other functions.

(138) Examples of controller or processing unit components that may be employed in various embodiments of the present disclosure include, but are not limited to, conventional microprocessors, application specific integrated circuits (ASICs), and field-programmable gate arrays (FPGAs).

(139) In various implementations, a processor or controller or processing unit may be associated with one or more storage media such as volatile and non-volatile computer memory such as RAM, PROM, EPROM, and EEPROM. The storage media may be encoded with one or more programs that, when executed on one or more processors and/or controllers and/or processing units, perform the required functions. Various storage media may be fixed within a processor or controller or processing unit or may be transportable, such that the one or more programs stored thereon can be loaded into a processor or controller or processing unit.

(140) Other variations to the disclosed embodiments can be understood and effected by those skilled in the art in practicing the claimed invention, from a study of the drawings, the disclosure, and the appended claims. In the claims, the word “comprising” does not exclude other elements or steps, and the indefinite article “a” or “an” does not exclude a plurality. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage. Any reference signs in the claims should not be construed as limiting the scope.