ULTRASOUND IMAGING APPARATUS, SIGNAL PROCESSING METHOD, AND SIGNAL PROCESSING PROGRAM

20220414836 · 2022-12-29

    Inventors

    Cpc classification

    International classification

    Abstract

    The invention provides an ultrasound imaging apparatus capable of highly accurately extracting a blood flow in a fine blood vessel in a short time. N pieces of frame data is generated by receiving ultrasound waves reflected by a subject with a plurality of transducers. A correlation matrix is generated based on a vector in which data at a corresponding position of the frame data is arranged for N frames, and a singular value and a singular vector for each of N ranks are calculated. A first filter element is calculated based on a variance between data at a corresponding position zx among a plurality of blood flow component frame data obtained by multiplying a plurality of the frame data by singular vectors at a threshold rank k or more. The second filter element is calculated based on the tissue component frame data obtained by multiplying the frame data by a singular vector at a rank 1. The frame data is weighted by the first filter element and/or the second filter element to generate a clutter reducing image.

    Claims

    1. An ultrasound imaging apparatus comprising: a data collection unit that receives a plurality of receive signals obtained by receiving, with an array in which a plurality of transducers are arranged in an x direction, ultrasound waves reflected sequentially in a depth z direction of a subject to which the ultrasound waves are transmitted, that repeats processing of generating frame data by arranging the receive signals on a zx plane, and that generates the frame data for N frames; a matrix conversion unit that generates a correlation matrix based on a vector in which data at a corresponding position zx of the frame data is arranged for N frames; a matrix analysis unit that performs singular value decomposition of the correlation matrix and that calculates the singular value and a singular vector for each of N ranks of the receive signals; a filter element generation unit that generates at least one of a first filter element and a second filter element as a filter element; and an image processing unit that uses the filter element generated by the filter element generation unit to weight the frame data or a pixel value of an image generated based on the frame data and generate a clutter reducing image, wherein the first filter element is calculated based on a variance between data at the corresponding position zx among a plurality of blood flow component frame data obtained by multiplying a plurality of the frame data by a plurality of singular vectors in a predetermined rank range of a preset threshold rank k or more, respectively and the second filter element is calculated based on tissue component frame data obtained by multiplying one or more of the frame data by a singular vector at a predetermined rank less than the threshold rank k.

    2. The ultrasound imaging apparatus according to claim 1, wherein the first filter element is a reciprocal of a value of the variance.

    3. The ultrasound imaging apparatus according to claim 1, further comprising: an image reconstruction unit that reconstructs a blood flow image which is extracted a blood flow component, wherein the image reconstruction unit generates the blood flow image in which data obtained by averaging a plurality of blood flow component frame data for each position zx is used as a pixel value, the plurality of blood flow component frame data being obtained by using a plurality of singular vectors in a predetermined rank range of the threshold ranks k to N as a blood flow extraction filter and multiplying a plurality of the frame data by the plurality of singular vectors, and the image processing unit weights the pixel value of the blood flow image by at least one of the first filter element and the second filter element to generate the clutter reducing image.

    4. The ultrasound imaging apparatus according to claim 3, wherein the image processing unit generates a mask image in which a pixel value is a value obtained by multiplying the first filter element and the second filter element by a weighting coefficient and adding and averaging products thereof for each position zx, and generates the clutter reducing image by multiplying a pixel value of the mask image at a position corresponding to the pixel value of the blood flow image.

    5. The ultrasound imaging apparatus according to claim 1, wherein the filter element generation unit calculates the first filter element using a plurality of singular vectors in a rank range from the threshold rank k to a rank N as a blood flow component filter.

    6. The ultrasound imaging apparatus according to claim 1, wherein the filter element generation unit generates the second filter element using a singular vector at a rank 1 as the singular vector at the predetermined rank less than the threshold rank k.

    7. The ultrasound imaging apparatus according to claim 1, wherein the matrix conversion unit obtains correlation matrices for a plurality of positions zx, and obtains an average correlation matrix of the correlation matrices, and the matrix analysis unit calculates the singular value and the singular vector based on the average correlation matrix obtained by the matrix conversion unit.

    8. The ultrasound imaging apparatus according to claim 4, wherein a sum of a weight coefficient of the first filter element and a weight coefficient of the second filter element is 1.

    9. The ultrasound imaging apparatus according to claim 1, further comprising: a singular value analysis unit that analyzes the singular value and that calculates the threshold rank k.

    10. The ultrasound imaging apparatus according to claim 4, wherein the image processing unit calculates a similarity of the plurality of frame data, and sets the weight coefficient based on a calculation result.

    11. The ultrasound imaging apparatus according to claim 10, wherein the image processing unit calculates the similarity of the plurality of frame data by calculating a similarity of the tissue component frame data.

    12. The ultrasound imaging apparatus according to claim 4, wherein the data collection unit calculates a similarity of the plurality of frame data, and sets, based on a calculation result, a number N of the frame data to be collected.

    13. A signal processing method comprising: a data collection step of receiving a plurality of receive signals obtained by receiving, with an array in which a plurality of transducers are arranged in an x direction, ultrasound waves reflected sequentially in a depth z direction of a subject to which the ultrasound waves are transmitted, repeating processing of generating frame data by arranging the receive signals on a zx plane, and generating the frame data for N frames; a matrix conversion step of generating a correlation matrix based on a vector in which data at a corresponding position zx of the frame data is arranged for N frames; a matrix analysis step of performing singular value decomposition of the correlation matrix and calculating the singular value and a singular vector for each of N ranks of the receive signals; a filter element generation step of generating at least one of a first filter element and a second filter element as a filter element; and an image processing step of using the filter element generated by the filter element generation unit to weight the frame data or a pixel value of an image generated based on the frame data and generate a clutter reducing image, wherein in the filter element generation step, the first filter element is calculated based on a variance between data at the corresponding position zx among a plurality of blood flow component frame data obtained by multiplying a plurality of the frame data by a plurality of singular vectors in a predetermined rank range of a preset threshold rank k or more, respectively and the second filter element is calculated based on tissue component frame data obtained by multiplying one or more of the frame data by a singular vector at a predetermined rank less than the threshold rank k.

    14. A signal processing program for causing a computer to execute: a data collection step of receiving a plurality of receive signals obtained by receiving, with an array in which a plurality of transducers are arranged in an x direction, ultrasound waves reflected sequentially in a depth z direction of a subject to which the ultrasound waves are transmitted, repeating processing of generating frame data by arranging the receive signals on a zx plane, and generating the frame data for N frames; a matrix conversion step of generating a correlation matrix based on a vector in which data at a corresponding position zx of the frame data is arranged for N frames; a matrix analysis step of decomposing the correlation matrix into a singular value and calculating the singular value and a singular vector for each of N ranks of the receive signals; a filter element generation step of generating at least one of a first filter element and a second filter element as a filter element; and an image processing step of using the filter element generated by the filter element generation unit to weight the frame data or a pixel value of an image generated based on the frame data and generate a clutter reducing image, wherein in the filter element generation step, the first filter element is calculated based on a variance between data at the corresponding position zx among a plurality of blood flow component frame data obtained by multiplying a plurality of the frame data by a plurality of singular vectors in a predetermined rank range of a preset threshold rank k or more, respectively and the second filter element is calculated based on tissue component frame data obtained by multiplying one or more of the frame data by a singular vector at a predetermined rank less than the threshold rank k.

    Description

    BRIEF DESCRIPTION OF THE DRAWINGS

    [0015] FIG. 1 is a block diagram of an ultrasound imaging apparatus according to a first embodiment;

    [0016] FIG. 2 is a flowchart showing processing executed by a blood flow image generation unit 30 of the ultrasound imaging apparatus according to the first embodiment;

    [0017] FIG. 3 is a flowchart illustrating the processing in FIG. 2 with figures and graphs;

    [0018] FIG. 4 is a view showing an example of a display screen of the ultrasound imaging apparatus according to the first embodiment;

    [0019] FIG. 5 is a graph showing a relationship between a correlation value (similarity) of a plurality of frame data calculated by an ultrasound imaging apparatus and a coefficient α of a second filter element according to a second embodiment;

    [0020] FIG. 6 is a flowchart showing processing executed by the blood flow image generation unit 30 of the ultrasound imaging apparatus according to the second embodiment;

    [0021] FIGS. 7A and 7B are views showing examples of a display screen of the ultrasound imaging apparatus according to the second embodiment;

    [0022] FIG. 8 is a flowchart showing another example of the processing executed by the blood flow image generation unit 30 of the ultrasound imaging apparatus according to the second embodiment;

    [0023] FIG. 9 is a flowchart showing processing executed by the blood flow image generation unit 30 of an ultrasound imaging apparatus according to a third embodiment; and

    [0024] FIG. 10 is a graph showing a relationship between a correlation value (similarity) of a plurality of frame data calculated by the ultrasound imaging apparatus and the number of frame data to be collected according to the third embodiment.

    DESCRIPTION OF EMBODIMENTS

    [0025] Ultrasound imaging apparatuses according to embodiments of the invention will be described with reference to the drawings.

    First Embodiment

    [0026] An ultrasound imaging apparatus 100 according to a first embodiment will be described with reference to FIGS. to 3. FIG. 1 is a diagram showing a schematic configuration of the ultrasound imaging apparatus according to the first embodiment, and FIG. 2 is a flowchart showing processing executed by a blood flow image generation unit of the ultrasound imaging apparatus. FIG. 3 is a flowchart illustrating the processing in FIG. 2 with figures and graphs.

    [0027] The configuration of the ultrasound imaging apparatus 100 according to the first embodiment will be described with reference to FIG. 1. The ultrasound imaging apparatus 100 is connected to a probe 10, an external input device 80 that receives an instruction from an outside, and a display device 70. The probe 10 includes an array in which ultrasound transducers are arranged in an x direction. Each transducer includes a piezoelectric element and the like.

    [0028] The ultrasound imaging apparatus 100 includes a transmission and receive control unit 20, a blood flow image generation unit 30, a memory 40, a B mode image generation unit 50, and a display processing unit 60.

    [0029] The transmission and receive control unit 20 includes a transmission beamformer 21, a receive beamformer 22, and an orthogonal detector (not shown). The transmission beamformer 21 generates transmission signals, outputs the transmission signals to the transducers of the probe 10, and transmits the transmission signals in a depth direction (z direction) of a subject 1. The receive beamformer 22 receives receive signals that are output by the transducers of the probe 10 after the transducers receive ultrasound waves, for example, reflected by the subject 1, executes predetermined phasing processing, and generates receive beam data. The orthogonal detector orthogonally detects the receive beam data and generates beam data as a complex signal.

    [0030] The blood flow image generation unit 30 includes a data collection unit 31, a matrix conversion unit 32, a matrix analysis unit 33, a singular value analysis unit 34, a filter element generation unit 35, a mask image generation unit 36, and an image reconstruction unit 37. The mask image generation unit 36 and the image reconstruction unit 37 constitute an image processing unit 38.

    [0031] The blood flow image generation unit 30 is implemented by a computer including a processor such as a central processing unit (CPU) or a graphics processing unit (GPU) and a memory, and the processor reads and executes a program stored in the memory in advance to implement functions of the units 31 to 38 by software. The blood flow image generation unit 30 may be partially or wholly implemented by hardware. For example, a custom IC such as an application specific integrated circuit (ASIC) or a programmable IC such as a field-programmable gate array (FPGA) may be used to design a circuit so as to implement the functions of the units 31 to 38.

    [0032] Hereinafter, operations of the units of the ultrasound imaging apparatus 100 according to the present embodiment will be described.

    [0033] The transmission beamformer 21 outputs the transmission signals to which a predetermined time delay is applied to the transducers of the probe 10. Accordingly, the transducers of the probe 10 irradiate, with the ultrasound waves, an organ tissue in a living body (subject 1) to be imaged by electroacoustic conversion or the like based on a piezoelectric effect. The ultrasound waves propagate in the living body so as to focus at a predetermined depth. In a process of the propagation, the ultrasound waves repeat reflection at interfaces having different acoustic impedances.

    [0034] The reflected waves (ultrasound waves) reach the transducers of the probe 10 again, are received, and are converted into the receive signals (electrical signals). The receive beamformer 22 executes the phasing processing on the receive signals so as to return the time delay set at the time of transmission, and generates, in the x direction, a plurality of receive beam data in the depth z direction. The receive signals after the phasing processing is sent to the blood flow image generation unit 30. However, as long as the receive signals treated in the invention are signals obtained by the reflection from an imaged target, a transmission and reception method thereof is not particularly limited, and for example, a signal obtained by transmitting a plane wave to an imaged target without applying a time delay to the transducers of the probe 10 may be used.

    [0035] When the blood flow image generation unit 30 generates a blood flow image, the ultrasound waves are repeatedly transmitted and received N times to and from the same imaging surface of the subject 1.

    [0036] Next, processing executed by the units 31 to 38 of the blood flow image generation unit 30 will be described with reference to the flowcharts in FIGS. 2 and 3.

    Step 201

    [0037] The data collection unit 31 receives a plurality of receive beam data after executing the phasing by the receive beamformer 22, arranges the received receive beam data on a zx plane, and generates frame data 301. The data collection unit 31 receives a signal each time the ultrasound waves are transmitted and received N times by the probe 10 in the same imaging range, and acquires the frame data 301 for N frames. The data collection unit 31 stores time-series (N pieces of) frame data 301 in the memory 40 in the form of an in-phase/quadrature-phase (IQ) signal. Accordingly, as shown in step 201 in FIG. 3, three-dimensional data 300 arranged in x, z, and N directions is stored in the memory 40.

    [0038] A quantity of the ultrasound waves reflected by a blood flow in the subject 1 is small, and most of the ultrasound waves are reflected at an interface of the structure in the subject 1. Therefore, most of the frame data 301 is a background component (clutter component) reflected by the structure, and a proportion of the blood flow component is fairly small. Therefore, the clutter component of the frame data 301 is removed and reduced by the following processing, and the blood flow component is extracted.

    Step 202

    [0039] The matrix conversion unit 32 extracts data u.sub.zxi of one pixel (position zx) of the frame data 301 from corresponding pixels (position zx) of the N pieces of frame data 301, and generates a vector u.sub.zx as indicated in an equation (1) by arranging the data for N frames. Here, i denotes a frame data number in the N direction. The vector u.sub.zx is ensemble data of singular value decomposition.

    [0040] Further, a correlation matrix R.sub.zx is generated using the vector u.sub.zx according to an equation (2).


    μ.sub.zx=[μ.sub.zx1, . . . ,μ.sub.zxN].sup.T . . .  (1)


    R.sub.zx=μ.sub.zxμ.sup.H.sub.zx . . .  (2)

    Step 203

    [0041] The matrix conversion unit 32 obtains the correlation matrix R.sub.zx for each of a plurality of pixels (for example, all pixels), averages the obtained correlation matrices according to an equation (3), and calculates an average correlation matrix R.sup.− (Here, “−” in R.sup.− denotes “−” above the letter R in the equation (3)).

    [00001] R _ = 1 XZ .Math. z = 1 Z .Math. x = 1 X R zx ( 3 )

    Step 204

    [0042] The matrix analysis unit 33 calculates, based on the average correlation matrix R.sup.− obtained by the matrix conversion unit 32, a singular value Ai and a singular vector w.sub.i (i=1, 2, . . . , N) for each rank i, that is, the first principal component to the Nth principal component according to an equation (4). (Here, the rank i denotes the number of each principal component obtained according to a principal component analysis based on the singular value decomposition, the first principal component corresponding to a higher order is called a rank 1 principal component, and the Nth principal component is called a rank N principal component). The equation (4) is a calculation process for the principal component analysis, but when the frame data 301 is a square matrix, eigenvalue decomposition may be used as an alternative method.

    [00002] R _ = W Λ W H = .Math. i = 1 N λ i w i w i H ( 4 )

    Steps 205 and 206

    [0043] Next, the singular value analysis unit 34 determines a threshold rank for separating a principal component of the blood flow component and a principal component of the clutter component by utilizing the fact that the blood flow component contained in the receive signals is significantly lower than the clutter component.

    [0044] The singular value analysis unit 34 calculates the singular value λ.sub.i (a diagonal component of a matrix A) of each principal component calculated by the singular value decomposition. Specifically, λ.sub.i/Σλ.sub.i is obtained by dividing the singular value λ.sub.i by a total addition number (Σλ.sub.i) of the singular value λ.sub.i (step 204). λ.sub.i/Σλ.sub.i indicates how much each principal component contributes to reproduction of original data, that is, the information dominance ratio.

    [0045] A graph 310 having λ.sub.i/Σλ, as a vertical axis and the rank i (i=1 to N) as a horizontal axis is generated as shown in step 205 in FIG. 3. Subsequently, a rank (rank threshold value k) of a principal component corresponding to a predetermined threshold value of λ.sub.i/Σλ.sub.i is selected. An example of the threshold value of λ.sub.i/Σλ.sub.i is 0.01, which follows that an amount of energy of the blood flow component with respect to a tissue component is approximately 1%.

    [0046] A preset value may be used as the rank threshold value k.

    Steps 207 and 208

    [0047] The singular vector (w.sub.i, i=k to N) below a rank threshold value (ranks k to N) can serve as a filter that reduces the clutter component contained in the receive signals and that extracts the blood flow component. The image reconstruction unit 37 uses the singular vector to reconstruct the blood flow image which is extracted the blood flow component.

    [0048] Specifically, the image reconstruction unit 37 generates a sum of products of a plurality of singular vectors and singular values in a predetermined rank range of the threshold ranks k to N as a blood flow extraction filter P.sub.k according to an equation (5) (step 207).

    [0049] Next, the image reconstruction unit 37 obtains blood flow component frame data U.sup.∧ by multiplying the plurality of frame data 301 (for example, three-dimensional data U which is all of frame data 301) by the blood flow extraction filter P.sub.k according to equations (6) and (7). (Here, “∧” in U.sup.∧ denotes “.sup.∧” above the letter U in the equation (6)).

    [0050] Further, the image reconstruction unit 37 generates a blood flow image U.sub.avg by adding and averaging the obtained three-dimensional data (N pieces of frame data) U.sup.∧ for each pixel (position zx) in a time axis direction (N direction) according to an equation (8) (step 208).

    [00003] P k = W Λ ( k ) W H .Math. = .Math. i = k N λ i w i w i H ( 5 ) U ^ = P k U ( 6 ) U = [ .Math. .Math. u zx .Math. .Math. ] ( 7 ) U avg = avg [ U ^ ] ( 8 )

    Steps 209 to 211

    [0051] The filter element generation unit 35 generates at least one of a first filter element and a second filter element as a filter element. In the present embodiment, both the first filter element and the second filter element are calculated.

    First Filter Element

    [0052] The first filter element is calculated based on a variance between data at the corresponding position zx among a plurality of blood flow component frame data. The plurality of blood flow component frame data are obtained by multiplying the plurality of frame data 301 by a plurality of singular vectors (w.sub.i, i=k to N) in a predetermined rank range of the threshold rank k or more (that is, a rank on a lower order side including the rank k) (step 209). Here, the first filter element is a reciprocal of the value of the variance.

    [0053] Specifically, a method for calculating the first filter element will be described. The filter element generation unit 35 obtains the blood flow extraction filter P.sub.k according to the equation (5) using the singular vector (w.sub.i, i=k to N) included in the rank threshold value k or less, and obtains N pieces of blood flow component frame data U.sup.∧ by multiplying the plurality of frame data 301 (N pieces of three-dimensional data U which are all of the frame data 301) by the blood flow extraction filter P.sub.k according to the equation (6). The filter element generation unit 35 calculates a variance (var [U.sup.∧]) in the time axis direction (N direction) for corresponding pixels (position zx) of the N pieces of blood flow component frame data U.sup.∧ according to the equation (9), and uses a reciprocal thereof as a first filter element U.sub.var. (In the equation (9), var [ ] denotes a function for calculating a brightness variance in the time axis direction).


    U.sub.var=var[.Math.].sup.−1 . . .  (9)

    var

    Second Filter Element

    [0054] The second filter element is calculated based on tissue component frame data obtained by multiplying one or more of frame data 301 by the singular vector w.sub.i at a predetermined rank less than the threshold rank k (steps 210 and 211). The filter element generation unit 35 uses a singular vector at a rank 1 as the singular vector at the predetermined rank less than the threshold rank k. The second filter element is a pixel value of the tissue component frame data.

    [0055] Specifically, a method for calculating the second filter element will be described. The filter element generation unit 35 generates a clutter extraction filter Pi according to an equation (10) based on a product of a singular vector w.sub.1 and a singular value λ.sub.1 of the first principal component from a result of the principal component analysis (step 210). The filter element generation unit 35 multiplies the clutter extraction filter Pi by the N pieces of frame data 301 (three-dimensional data U), and calculates N pieces of tissue component frame data (three-dimensional data) U.sub.1.sup.∧ according to an equation (11).

    [0056] The filter element generation unit 35 obtains frame data U.sub.1 indicating a distribution of a brightness centered on the clutter component by adding and averaging pixels corresponding to the obtained N pieces of tissue component frame data (three-dimensional data) U.sub.1.sup.∧ in time-series in the time axis direction (N direction) according to an equation (12), and uses the frame data U.sub.1 as the second filter element. (In the equation (12), avg [ ] denotes a function for calculating an average in the time axis direction (N direction).

    [00004] P 1 = W Λ ( 1 ) W H = .Math. i = 1 1 λ i w i w i H ( 10 ) U 1 ^ = P 1 U ( 11 ) U 1 = avg [ U 1 ^ ] ( 12 )

    [0057] The image processing unit 38 weights the frame data 301 or a pixel value of an image generated based on the frame data by at least one of the first filter element U.sub.var and the second filter element U.sub.1 to generate a clutter reducing image. Here, the image processing unit 38 weights the pixel value of the blood flow image U.sub.avg obtained in step 208 by the first filter element U.sub.var and the second filter element U.sub.1 to generate the clutter reducing image.

    [0058] Specifically, the mask image generation unit 36 executes weighting and adding by multiplying the first filter element U.sub.var and the second filter element U.sub.1 by coefficients β and α (an addition result of the weight coefficients α and β is 1) according to an equation (13), and generates a mask image MASK (two-dimensional coefficient matrix (mask matrix)) having a reciprocal value of the weighting-added value as the pixel value (step 212). For a relationship between α and β, by setting the addition result thereof to 1 as described above, an advantage is present that variable parameters can be unified and an adjustment executed by a user can be simplified. However, when priority is given to an image quality, it may be effective to treat α and β as independent variables, and α design that can be switched according to a purpose of an operator is effective.


    MASK.sub.(α+β=1)=AVG(αU.sub.1+βU.sub.var).sup.−1 . . .  (13)

    [0059] The first filter element U.sub.var is a reciprocal of a brightness variance of the blood flow component frame data in the time direction (N direction) as described above. In general, the blood flow component having a three-dimensional flow has a large temporal change in the brightness as compared with a movement of the organ tissue. Therefore, by using the reciprocal of the brightness variance as a pixel value of a mask, the mask image MASK having a high numerical value (pixel value) can be generated in pixels of a tissue region of the subject 1.

    [0060] On the other hand, the second filter element U.sub.1 is a brightness average of the clutter component (tissue component) frame data (three-dimensional data) U.sub.1.sup.∧ extracted according to the singular vector of the first principal component. As compared to brightness data of the frame data of the receive signals, the clutter component is dominated by signals that are spatially and temporally stable, and is less susceptible to a blood flow and a noise. Therefore, by using a reciprocal of the second filter element U.sub.1 as the pixel value of the mask, the mask image MASK can be generated from which a boundary and a high brightness region of the organ tissue can be selectively extracted.

    [0061] In other words, the mask image generation unit 36 uses the first filter element and the second filter element, both of which show high numerical values in the tissue region that is a clutter source, and sets reciprocals thereof to mask pixel values. Therefore, the mask image MASK having an effect of reducing the clutter can be generated (step 213).

    [0062] The image reconstruction unit 37 multiplies the blood flow image U.sub.avg generated in step 208 by the mask image MASK according to an equation (14) and executes image reconstruction to generate a highly visible blood flow image U.sub.MASK in which a residual clutter is small and a blood flow is emphasized (step 213).


    U.sub.MASK=U.sub.avg×MASK . . .  (14)

    [0063] The image reconstruction unit 37 can generate the blood flow image U.sub.MASK in which the residual clutter is small and a blood flow region is emphasized. The image reconstruction unit 37 outputs the blood flow image U.sub.MASK to the display processing unit 60.

    [0064] On the other hand, the B mode image generation unit receives the receive beam data from the receive beamformer 22, generates a B mode image, and outputs the B mode image to the display processing unit 60.

    [0065] For example, as shown in FIG. 4, the display processing unit 60 displays the B mode image and the blood flow image U.sub.MASK in which the blood flow region is emphasized in regions 401 and 402 of a screen of the display device 70, respectively. At this time, for comparison, the blood flow image U.sub.avg generated by the image reconstruction unit 37 in step 208 may be displayed together with a region 403.

    [0066] The mask image generation unit 36 may accept, from the user, the values of the coefficients β and α (α+β=1) of the first filter element U.sub.var and the second filter element U.sub.1, which are used during generating the mask image MASK according to the equation (13), on the display screen of the display device 70. For example, the display processing unit 60 causes a bar 404 in FIG. 4 to be displayed on the display screen. The user sets a ratio of the coefficient α and the coefficient β by moving a mark 405 on the bar 404 left and right by the external input device 80 such as a mouse. The mask image generation unit 36 generates, in step 212, the mask image MASK using the set ratio of the coefficient α and the coefficient β.

    [0067] A first feature according to the present embodiment is that the second filter element U.sub.1 using the first principal component which is the tissue component (clutter component) is used. Accordingly, not only a low brightness region of the organ tissue but also the clutter component remaining in the high brightness region can be reduced by the mask image.

    [0068] That is, since the second filter element U.sub.1 is an image at a rank 1 (first principal component), the second filter element U.sub.1 is an image of only a tissue (background) and does not include the blood flow component or an electrical noise. Therefore, the second filter element U.sub.1 is more suitable than an original image (frame data) for removing components having a high brightness such as blood vessel walls contained in the blood flow image.

    [0069] A second feature according to the present embodiment is that the first filter element U.sub.var using dispersion of the blood flow image U.sub.avg in the time direction is used. By adjusting a brightness of the blood flow image U.sub.avg with the mask image according to the temporal change in the brightness of the blood flow image U.sub.avg, not only the clutter component can be reduced, but also the blood flow component can be expected to be enhanced.

    [0070] That is, since the first filter element U.sub.var is calculated using the N pieces of blood flow component frame data U.sup.∧ containing abundant blood flow components, a blood flow can be effectively extracted.

    [0071] In the present embodiment, both the generation of the blood flow image U.sub.avg and the generation of the first and second filter elements are performed in the principal component analysis. Therefore, in order to reduce the clutter component, calculation efficiency is higher than that when processing different from the principal component analysis is performed, and a blood flow in a fine blood vessel can be highly accurately extracted in a short time. No other processing circuits are required, and α device configuration can be simplified.

    Second Embodiment

    [0072] An ultrasound imaging apparatus according to a second embodiment will be described. The ultrasound imaging apparatus according to the second embodiment has the same configuration as the apparatus according to the first embodiment, but further has a function of appropriately setting coefficients β and α (α+β=1) to be multiplied by the first filter element U.sub.var and the second filter element U.sub.1. That is, the image processing unit 38 calculates a similarity of the plurality of frame data 301, and sets the weighting coefficients α and β based on a calculation result.

    [0073] When a correlation (similarity) between the N pieces of frame data 301 for executing a principal component analysis is high, it is appropriate that the coefficients β and α are both around 0.5. This is because a clutter component and a blood flow component can be effectively separated according to the principal component analysis when the correlation (similarity) between the N pieces of frame data 301 is high, and a brightness of the clutter component is also high in the calculation result of the second filter element U.sub.1. Accordingly, the mask image MASK obtained by taking a reciprocal after weighting and adding the first filter element U.sub.var and the second filter element U.sub.1 is generated, and is multiplied by the blood flow image U.sub.avg, thereby achieving a clutter reducing effect.

    [0074] On the other hand, when the correlation of the N pieces of frame data 301 is low, it is assumed that the clutter component is present across a plurality of principal components and a brightness of the clutter component of the second filter element U.sub.1 is low. When the mask image MASK is configured in step 212 in this state, a phenomenon may occur in which the clutter component of the blood flow image U.sub.avg is rather emphasized by multiplying the blood flow image MASK by the blood flow image U.sub.avg.

    [0075] In order to avoid this phenomenon, when the correlation between the N pieces of frame data 301 is low, a method of setting a value of the coefficient α to be low to reduce a contribution ratio of the second filter element U.sub.1 to the mask image MASK is effective.

    [0076] For example, as shown in FIG. 5, a linear graph 501 and curve graphs 502 and 503 are set such that the coefficient α of the second filter element U.sub.1 becomes smaller as the correlation value of the N pieces of frame data 301 becomes lower, and are stored in the memory 40.

    [0077] As shown in a flow in FIG. 6, the mask image generation unit 36 executes steps 214 and 215 to calculate the correlation value of the N pieces of frame data 301, and obtains the coefficient α corresponding to the calculated correlation value based on any of the graphs 501, 502, and 503 in FIG. 5. Then, in step 212, the second filter element U.sub.1 and the first filter element U.sub.var are weighted using the obtained coefficient α and β (=1−α) according to the equation (13) to obtain the mask image MASK.

    [0078] Accordingly, the appropriate coefficient α can be automatically set in the ultrasound imaging apparatus.

    [0079] As shown in FIGS. 7A and 7B, a graph 1001 showing a temporal change (stability of data) of the correlation value calculated in step 214 may be displayed on the display device 70 together with the blood flow image U.sub.avg after reducing a residual clutter generated in step 213. FIG. 7A shows a case in which the correlation value is maintained high and the stability of the data is high, and the coefficient α is set to a large value close to 0.5. On the other hand, FIG. 7B shows a case in which the correlation value is maintained low and the stability of the data is low, and a ratio of the coefficient α is set to be small.

    [0080] A calculation of the correlation value of the N pieces of frame data 301 in step 214 uses a generally known similarity calculation method such as a cross-correlation calculation.

    [0081] In the flow in FIG. 6, steps other than the above steps is the same as those in the flow in FIG. 2 according to the first embodiment, and thus a description thereof will be omitted.

    [0082] As shown in a flow in FIG. 8, the mask image generation unit 36 can also determine the similarity of the N pieces of frame data 301 using the tissue component frame data (U.sub.1.sup.∧) obtained using a first principal component.

    [0083] The tissue component frame data is a component (three-dimensional data) mainly including a tissue structure extracted according to a singular vector of the first principal component. As described above, signals reflected from a living body mainly include a component from organ tissues, and a component from a blood flow is 10% to 1% or less. That is, the similarity of the data is appropriate for judging mainly the component of the organ tissue.

    [0084] Accordingly, the mask image generation unit 36 executes steps 216 and 217 before step 212 to calculate the correlation value of the tissue component frame data, and obtains the coefficient α corresponding to the calculated correlation value based on the graph in FIG. 5.

    [0085] By calculating the correlation value of the N pieces of frame data 301 using the tissue component frame data, an influence of a tissue movement and a noise that affect the principal component analysis can be eliminated, and the correlation can be determined more effectively. Accordingly, the appropriate coefficient α can be automatically set in the ultrasound imaging apparatus. Here, α and β are treated as dependent variables (α+β=1). However, this limitation is a setting that focuses on convenience of an operation, and does not limit the treating of α and β as independent variables (that is, 0<=α<=1 and 0<=β<=1) while maintaining the described apparatus form.

    [0086] In the flow in FIG. 8, steps other than the above steps is the same as those in the flow in FIG. 2 according to the first embodiment, and thus a description thereof will be omitted.

    [0087] A configuration, an operation, and an effect of the ultrasound imaging apparatus according to the second embodiment other than those described above are the same as those according to the first embodiment, and thus a description thereof will be omitted.

    Third Embodiment

    [0088] An ultrasound imaging apparatus according to a third embodiment will be described. The ultrasound imaging apparatus according to the third embodiment has the same configuration as the apparatus according to the first embodiment, but further has a function of determining a correlation (similarity) between the N pieces of frame data 301 and adjusting an acquired number of the data.

    [0089] When the number N of the frame data 301 used for a principal component analysis is large in the matrix conversion unit 32, the matrix analysis unit 33, the singular value analysis unit 34, and the filter element generation unit 35, the number of distinguished principal components increases and accuracy of separation between α clutter and a blood flow increases. However, a calculation load in the blood flow image generation unit 30 is heavy, which may lead to a decrease in a frame rate.

    [0090] Therefore, in the third embodiment, when the correlation between the frame data 301 is sufficiently high, both effects of reducing the clutter and improving the frame rate are achieved by reducing the acquired number N.

    [0091] Specifically, as shown in a flow in FIG. 9, step 200 of setting the number N of the frame data 301 to be collected is executed before step 201 of collecting the N pieces of frame data 301.

    [0092] In step 200, a correlation value of tissue component frame data calculated in previous step 216 is received, and the appropriate number N of frames is set according to, for example, a graph 901, 902 or 903 showing a relationship between a predetermined correlation value and the number of frames as shown in FIG. 10. The correlation value of the tissue component frame data corresponds to a correlation value of the N pieces of frame data 301 as described in the second embodiment.

    [0093] As shown in FIG. 10, the linear graph 901 and the curve graphs 902 and 903 are set such that the higher the correlation value of the N pieces of frame data 301, the smaller the number N of frames to be acquired. These graphs 901 to 903 are stored in the memory 40.

    [0094] By using the flow in FIG. 9, the number N of the frame data 301 used for the current principal component analysis can be set to an appropriate number using the correlation value of the tissue component frame data calculated in previous step 216.

    [0095] As a method for calculating the correlation value of the N pieces of frame data 301, in addition to the method of calculating the correlation value of the tissue component frame data, the correlation value may be calculated directly based on the frame data 301 in step 214 of the flow in FIG. 6.

    [0096] If the value of the appropriate number N of frames changes significantly each time the principal component analysis is executed, a brightness flicker occurs at a timing of switching a displayed blood flow image, which may impair visibility as a moving image. In order to prevent the brightness flicker, it is effective to provide a limiting measure in which, for example, a fluctuation of the number of data between consecutive frames is 2 frames or less.

    [0097] A configuration, an operation, and an effect of the ultrasound imaging apparatus according to the third embodiment other than those described above are the same as those according to the first embodiment and the second embodiment, and thus a description thereof will be omitted.