SYSTEM FOR IMAGE PROCESSING

20220378515 · 2022-12-01

    Inventors

    Cpc classification

    International classification

    Abstract

    The invention is directed to a system (100) for image processing, which is capable of improving an usability of an imaging device during an intervention.

    Claims

    1. A system for image processing, comprising: an identification unit, and a compliance unit, wherein the identification unit is configured to receive image data representing an intervention and identify an interventional device used in the intervention and an anatomical region of the intervention, based on the image data, wherein the identification unit is further configured to determine an intervention focus on the basis of the identified device and the identified anatomical region, wherein the identification unit is further configured to determine an orientation parameter indicative for an orientation of an imaging device generating the image data; wherein the compliance unit is further configured to receiving a guideline corresponding to a type of the intervention; wherein the compliance unit is further configured to determine, based on the guideline, a guideline position parameter corresponding to the intervention focus and to r compare the determined guideline position parameter and the determined orientation parameter and to calculate a deviation between the guideline position parameter and the orientation parameter; wherein the compliance unit is configured to generate a signal indicative of a positioning discrepancy if the calculated deviation exceeds a pre-defined threshold.

    2. System according to claim 1, wherein the compliance unit is configured for calculating an optimized orientation parameter for the orientation of the imaging device such that the deviation is minimized.

    3. System according to claim 2, wherein the compliance unit is configured for generating a command signal based on the optimized orientation parameter, and wherein the command signal comprises commands instructing a user to adjust the imaging device, which will minimize the deviation.

    4. System according to claim 1, wherein the orientation parameter and/or the optimized orientation parameter is a geometrical parameter and/or an angulation parameter of the imaging device and/or of an operating table of the imaging device.

    5. System according to claim 1, wherein the identification unit is configured for identifying a position and/or an orientation of the device, used in the intervention and depicted in the medical image data, and wherein the determination of the intervention focus is further based on the position and/or the orientation of the device.

    6. System according to claim 1, wherein the identification unit is configured for identifying an anatomical element within the anatomical region depicted in the image data and/or a report and log data element, wherein the identification unit is configured for determining a position and/or an orientation of the anatomical element in the image data and/or the report and log data element, wherein the determination of the intervention focus is further based on the position and/or the orientation of the anatomical element depicted in the image data.

    7. System according to claim 1, wherein the identification unit is configured for correlating the device with the anatomical region by a registration of a fluoroscopic image data of the intervention.

    8. System according to claim 1, further comprising: a decision unit configured to statistically evaluate a result of the comparison between the pre-defined threshold with the orientation parameter.

    9. System according to claim 1, wherein the decision unit is configured for dynamically adapting the intervention focus throughout an interventional procedure.

    10. System according to claim 1, wherein the system comprises an artificial-intelligence (AI) module, wherein the AI module is configured for identifying the type of the intervention based on analyzing the image data and/or analyzing logging information.

    11. System according to claim 10, wherein the compliance unit is configured for adapting the pre-defined threshold based on the determined type of the intervention.

    12. Computer-implemented method, the method comprising the steps of: receiving image data representing an intervention; identifying an interventional device used in the intervention and an anatomical region of the intervention, based on the image data, determining an intervention focus on the basis of the identified device and the identified anatomical region, determining an actual orientation parameter indicative for an orientation of an imaging device, when generating the image data, receiving a guideline corresponding to a type of intervention, determining, based on the guideline, a guideline position parameter corresponding to the intervention focus, comparing the determined guideline position parameter and the determined orientation parameter d, calculating a deviation between the guideline position parameter and the orientation parameter, and, generating a signal indicative of a positioning discrepancy if the calculated deviation exceeds a pre-defined threshold.

    13. Computer program, which when executed on a processor, instructs the processor to perform the method according to claim 12.

    14. Imaging device, comprising: a system and/or a computer readable medium on which a computer program according to claim 13 is stored.

    Description

    BRIEF DESCRIPTION OF THE DRAWINGS

    [0077] FIG. 1 shows a system for imaging processing according to an embodiment.

    [0078] FIG. 2 shows a medical imaging device according to an embodiment.

    [0079] FIG. 3 shows a schematic intervention focus according to an embodiment.

    [0080] FIG. 4 shows a flow-chart diagram illustrating the method according to an embodiment.

    DETAILED DESCRIPTION OF EMBODIMENTS

    [0081] FIG. 1 shows a system 100 for imaging processing, wherein the system 100 comprises an identification unit 102 and a compliance unit 104. The identification unit 102 is configured for determining an intervention focus 200. The determined intervention focus 200 specifies which intervention is performed in an image data and specifies which anatomical region 202 is treated by the intervention in the image data. Furthermore, the identification unit 102 is configured for determining an orientation parameter indicative for an orientation 302 of an imaging device 300 during generation of the image data.

    [0082] In addition, the compliance unit 104 is configured to receive a guideline corresponding to a type of intervention. Based on the guideline, a guideline position parameter corresponding to the intervention focus is determined, which parameter is compared with the current orientation parameter. A deviation between the guideline position parameter and the orientation parameter is calculated.

    [0083] The compliance unit 104 is further configured to receive, determine or store a predefined threshold for the orientation parameter specific for the received intervention focus 200. In addition, the compliance unit 104 is configured for comparing the calculated deviation and the predefined threshold.

    [0084] In certain embodiments, furthermore, the compliance unit 104 is optionally configured for calculating an optimized orientation parameter for amending the orientation 302 of the imaging device 300. Furthermore, the compliance unit 104 may be configured for generating a control signal for changing the orientation 302 of the imaging device 300 based on the calculated optimized orientation parameter.

    [0085] The advantage of this embodiment is that the positioning of a imaging device 300 can be improved during intervention. By providing a solution for such a dynamic process, the usability of an imaging device 300 during an intervention can be improved. Furthermore, this embodiment can comprise the advantage that with the help of the supervising of the intervention, a controlling factor can be implemented, which is further increasing the reliability of the intervention. In addition, predefined guidelines can be used in order to increase the usability of the imaging device 300 during an intervention since the system is capable of comparing the current situation of the intervention with the predefined guidelines.

    [0086] As can be seen in FIG. 1, the system 100 comprises the identification unit 102. The identification unit 102 can be configured for determining an intervention focus 200, for example by determining an intervention focus 200 based on medical image data

    [0087] Furthermore, the identification unit 102 is configured to determine an orientation parameter indicative for an orientation 302 of an imaging device 300.

    [0088] In particular, the identification unit 102 is configured for determining the intervention focus 200 based on the image data. Furthermore, the identification unit 102 is configured for identifying a device 204 and the anatomical region 202 in the image data. Additionally, the identification unit 102 is configured for identifying a position 206 and/or the orientation 208 of the device 204 within the medical image data. Furthermore, the identification unit 102 is configured for identifying an anatomical element 210, for example a coronary artery in the medical image data. Furthermore, the identification unit 102 is configured to determine a position 212 and an orientation 214 of the anatomical element 210 in the image data and/or in a report and/or log data element.

    [0089] In addition, in certain embodiments the identification unit 102 is configured for correlating the device 204 with the anatomical region 202 with the help of the registration of fluoroscopy (idem) within the medical image data.

    [0090] In certain embodiment, the compliance unit 104 is configured for generating a control signal for changing the orientation 302 of the imaging device 300, in particular so as to minimize the deviation between the guideline position parameter and the actual orientation parameter. In addition, the compliance unit 104 may be configured for generating a command signal. The command signal comprises commands instructing a user 400 to adjust the position and/or the orientation of the imaging device 300 accordingly.

    [0091] Furthermore, the compliance unit 104 is configured for adapting the predefined threshold based on the determined type of the intervention. Alternatively or in addition, the predefined threshold may be obtained from the guideline along with the guideline position parameter.

    [0092] Furthermore, the system 100 comprises a decision unit 106 which is configured to statistically evaluate a result of the comparison between the predefined threshold with the orientation parameter. In addition, the decision unit 106 is configured for dynamically adapting the intervention focus 200 throughout an interventional procedure. In addition, the system 100 comprises an artificial (AI) module 108. The AI module 108 is configured for identifying a type of the intervention based on the image data. Furthermore, the system 100 can comprise a user interface 110 which is configured for receiving a type of intervention or an intervention focus 200. In addition, the system 100 can comprise a logging unit 112 which is configured to record the intervention and in particular data and logs of the intervention focus 200 in order to create a report and log data element.

    [0093] FIG. 2 shows an imaging device 300 which comprises a system 100 as described before and hereinafter. In addition, the imaging device 300 comprises a computer-readable medium 600 on which a computer program as described before and herein later is stored. In addition, the imaging device can comprise an interface element 350 which is configured to interact with a user 400 of the imaging device 300. The interface element 350 can for example indicate a change of the orientation 302 of the imaging device 300. Furthermore, the imaging device can comprise the operating table 304, wherein the interface element 350 can output signals which indicate a change of orientation of the operating table 304.

    [0094] FIG. 3 shows an intervention focus 200 of an anatomical region 202. Within the intervention focus 200, a device 204 can be identified. The device 204 can be located within the image data according to a position 206. Furthermore, the orientation 208 of the device can be identified within the intervention focus 200. Furthermore, within the intervention focus 200 and/or the anatomical region 202, an anatomical element 210 can be identified. Furthermore, the position 212 of the anatomical element 210 can be assessed. In addition, the orientation 214 of the anatomical element 210 can be assessed with the help of the system 100.

    [0095] FIG. 4 shows a flow-chart diagram illustrating the method 500. The method 500 comprises the step S1 of determining an intervention focus from image data representing an intervention and particularly on the basis of an interventional device and/or anatomical region identified in said image data. Furthermore, the method 500 comprises a step S2 of determining an orientation parameter. Furthermore, the method 500 comprises the step S3 of comparing a guideline position parameter determined from a guideline corresponding to a type of intervention to the determined orientation parameter. In addition, the method 500 comprises the step S4 of calculating a deviation between the guideline position parameter and the orientation parameter. Furthermore, the method 500 comprises a step of generating S5 a signal indicative of a positioning discrepancy.

    [0096] Where an indefinite or definite article is used when referring to a singular noun, e.g. “a”, “an” or “the”, this includes the plurality of the noun unless something else is specifically stated. It is to be understood that the terms so used are interchangeable under appropriate circumstances and that the embodiments of the invention described here are capable of operating in other sequences than described or illustrated herein.

    LIST OF REFERENCE SIGNS

    [0097] 100—system [0098] 102—identification unit [0099] 104—compliance unit [0100] 106—decision unit [0101] 108—AI module [0102] 110—interface [0103] 112—logging unit [0104] 200—intervention focus [0105] 202—anatomical region [0106] 204—device [0107] 206—position [0108] 208—orientation [0109] 210—anatomical element [0110] 212—position [0111] 214—orientation [0112] 300—imaging device [0113] 302—orientation [0114] 304—operating table [0115] 350—interface [0116] 400—user [0117] 500—method [0118] 600—computer-readable medium [0119] S1—receiving an intervention focus [0120] S2—determining an orientation parameter [0121] S3—comparing [0122] S4—calculating an optimized orientation parameter [0123] S5—generating