METHODS AND SYSTEM FOR AUTONOMOUS VOLUMETRIC DENTAL IMAGE SEGMENTATION
20220012888 · 2022-01-13
Inventors
Cpc classification
G06T7/187
PHYSICS
International classification
A61C7/00
HUMAN NECESSITIES
Abstract
The present disclosure describes a system and methods for autonomous segmentation of volumetric dental images, such as those produced by an imaging system, The methods, implemented by the system, acquire a volume image of a patient and extract a volume of interest comprising patient dentition from the acquired volume image. A first plane is extended through maxillary portions of the patients jaw and a second plane through mandibular portions of the patients jaw. A maxillary sub-volume is generated from the volume of interest according to the first plane and a mandibular sub-volume from the volume of interest according to the second plane. Maximum intensity projection images are formed for each sub-volume and teeth are delineated from these images. Teeth are segmented within each sub-volume according to the tooth delineation for their respective sub-volume.
Claims
1. A method for tooth segmentation, comprising: acquiring a volume image of a patient; identifying a first plane extending through maxillary portions of the patient's jaw and a second plane extending through mandibular portions of the patient's jaw; and generating a maxillary sub-volume from the volume of interest according to the first plane and a mandibular sub-volume from the volume of interest according to the second plane.
2. The method of claim 1, wherein the method further comprises: forming, for each sub-volume, a maximum intensity projection image MIP from voxels of the corresponding sub-volume; delineating teeth from the MIP data to define tooth contour within each corresponding sub-volume; and segmenting and displaying teeth within each respective sub-volume according to the tooth delineation.
3. The method of claim 1, wherein the step of identifying the first or second plane comprises accepting operator input for positioning the plane with respect to the volume image.
4. The method of claim 1, wherein the method further comprises a step of extracting a volume of interest comprising patient dentition from the acquired volume image.
5. The method of claim 1, wherein the step of identifying the first or second plane comprises processing volume data to align the plane to tooth structure.
6. The method of claim 1, wherein the step of acquiring the volume image comprises acquiring cone-beam computed tomography image content.
7. The method of claim 2, wherein the step of delineating teeth from the MIP data comprises a step of forming a spline corresponding to the arrangement of teeth in the sub-volume, and a step of calculating distances to tooth boundaries for points along the spline.
8. The method of claim 2, wherein the step of segmenting teeth comprises using a level set method.
9. The method of claim 2, wherein the step of forming the MIP for the maxillary or mandibular sub-volume comprises a step of defining and using a normal to the corresponding first or second plane.
10. The method of claim 2, wherein the method further comprises a step of executing a random walk algorithm on the MIP data.
11. The method of claim 2, wherein the method further comprises a step of computing a medial axis for one or more teeth.
12. A method for tooth segmentation, the method comprising the steps of: acquiring a cone beam computed tomography volume image of a subject; accepting an operator instruction that defines a first plane extending through maxillary portions of the patient's jaw and a second plane extending through mandibular portions of the patient's jaw; generating a maxillary sub-volume from the volume of interest according to the first plane; and generating a mandibular sub-volume from the volume of interest according to the second plane.
13. The method of claim 1, wherein the method further comprises the steps of: generating, for each sub-volume, a 2D maximum intensity projection image from voxels of the corresponding sub-volume; delineating teeth from the 2D MIP data within each corresponding sub-volume; segmenting teeth within each respective sub-volume according to the tooth delineation; and computing and displaying cephalometric parameters for diagnosis using the tooth segmentation.
14. The method of claim 12, wherein the method further comprises a step of extracting a volume of interest from the acquired volume image, wherein the volume of interest comprises patient dentition.
15. The method of claim 12, wherein the step of forming the mandibular sub-volume comprises the steps of using the portion of the volume image on one side of the second plane, and adding connected portions of the volume image that lie between the first and second planes.
16. The method of claim 13, wherein the step of generating the 2D maximum intensity projection image comprises assessing voxel values aligned along a normal to the first or second plane.
17. The method of claim 13, wherein the step of delineating teeth from the 2D MIP data further comprises applying a random walk algorithm.
18. The method of claim 13, wherein the step of computing and displaying cephalometric parameters comprises a step of displaying a medial axis for one or more segmented teeth.
19. The method of claim 13, wherein the step of segmenting further comprises a step of identifying one or more false negative or false positive conditions.
20. The method of claim 19, wherein the method further comprises the steps of correcting for the false positive condition by generating an intermediate result using a region-driven level-set segmentation, and processing the generated intermediate result by applying an edge-driven level set segmentation.
21. The method of claim 19, wherein the method further comprises identifying a region within a slice having a level-set transition from another slice and applying erosion over the identified region.
22. The method of claim 13, wherein the step of segmenting further comprises applying a shrink or expand force to a level-set segmentation algorithm.
23. An imaging apparatus, comprising: an x-ray source and receiver configured to acquire a plurality of projection images of a patient; a processor configured to: (i) form a volume image of patient dentition from the acquired projection images; (ii) identify a first plane extending through maxillary portions of the patient's jaw and a second plane extending through mandibular portions of the patient's jaw according to operator instructions; (iii) generate a maxillary sub-volume from the volume of interest according to the first plane and a mandibular sub-volume from the volume of interest according to the second plane; (iv) form, for each sub-volume, a maximum intensity projection image MIP from voxels of the corresponding sub-volume; (v) delineate teeth from the MIP data to define tooth contour within each corresponding sub-volume; and (vi) segment and display teeth within each respective sub-volume according to the tooth delineation.
24. The apparatus of claim 23, wherein the x-ray source and receiver are part of a cone beam computed tomography system.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0011]
[0012]
[0013]
[0014]
[0015]
[0016]
[0017]
[0018]
[0019]
[0020]
[0021]
[0022]
[0023]
[0024]
[0025]
[0026]
[0027]
[0028]
[0029]
[0030]
[0031]
[0032]
[0033]
DETAILED DESCRIPTION OF EXAMPLE EMBODIMENTS
[0034] In the following detailed description of example embodiments of the present invention, reference is made to the drawings in which the same reference numerals are assigned to identical elements or steps in successive figures. It should be noted that these figures are provided to illustrate overall functions and relationships according to embodiments of the present invention and are not provided with intent to represent actual size or scale.
[0035] Where they are used in the context of the present disclosure, the terms “first”, “second”, and so on, do not necessarily denote any ordinal, sequential, or priority relation, but are simply used to more clearly distinguish one step, element, or set of elements from another, unless specified otherwise.
[0036] As used herein, the term “energizable” relates to a device or set of components that perform an indicated function upon receiving power and, optionally, upon receiving an enabling signal.
[0037] In the context of the present disclosure, the terms “viewer”, “operator”, and “user” are considered to be equivalent and refer to the viewing practitioner, technician, or other person who views and manipulates an image, such as a dental image, on a display monitor. An “operator instruction” or “viewer instruction” is obtained from explicit commands entered by the viewer, such as by clicking a button on a camera or by using a computer mouse or by touch screen or keyboard entry.
[0038] In the context of the present disclosure, the phrase “in signal communication” indicates that two or more devices and/or components are capable of communicating with each other via signals that travel over some type of signal path. Signal communication may be wired or wireless. The signals may be communication, power, data, or energy signals. The signal paths may include physical, electrical, magnetic, electromagnetic, optical, wired, and/or wireless connections between the first device and/or component and second device and/or component. The signal paths may also include additional devices and/or components between the first device and/or component and second device and/or component.
[0039] In the context of the present invention, the descriptive terms “object of interest” or “feature of interest” generally indicate an object such as a tooth or other object in the mouth.
[0040] The term “set”, as used herein, refers to a non-empty set, as the concept of a collection of elements or members of a set is widely understood in elementary mathematics. The term “subset”, unless otherwise explicitly stated, is generally used herein to refer to a non-empty proper subset, that is, to a subset of the larger set, having one or more members. For a set “S”, a subset may comprise the complete set “S”. A “proper subset” of set “S”, however, is strictly contained in set “S” and excludes at least one member of set “S”.
[0041] In the context of the present disclosure, the terms “pixel” and “voxel” may be used interchangeably to describe an individual digital image data element, that is, a single value representing a measured image signal intensity. Conventionally, an individual digital image data element is referred to as a voxel for 3-dimensional volume images and a pixel for 2-dimensional images. Volume images, such as those from CT or CBCT apparatus, are formed by obtaining multiple 2D images of pixels, taken at different relative angles, then combining the image data to form corresponding 3D voxels. For the purposes of the description herein, the terms voxel and pixel can generally be considered equivalent, describing an image elemental datum that is capable of having a range of numerical values. Voxels and pixels have the attributes of both spatial location and image data code value.
[0042] For general description and background on CT imaging, reference is hereby made to U.S. Pat. No. 8,670,521 entitled “Method for Generating an Intraoral Volume Image” by Bothorel et al., commonly assigned.
[0043] Overview of Dental CBCT Apparatus
[0044] The schematic diagram of
[0045] Volume Image Reconstruction from Multiple Projection Images
[0046] The schematic diagram of
[0047] CBCT imaging apparatus and the imaging algorithms used to obtain 3D volume images using such systems are well known in the diagnostic imaging art and are, therefore, not described in detail in the present application. Some exemplary methods and approaches for forming 3D volume images from the source 2D images, projection images that are obtained in operation of the CBCT imaging apparatus can be found, for example, in the teachings of U.S. Pat. No. 5,999,587 entitled “Method of and System for Cone-Beam Tomography Reconstruction” to Ning et al. and of U.S. Pat. No. 5,270,926 entitled “Method and Apparatus for Reconstructing a Three-Dimensional Computerized Tomography (CT) Image of an Object from Incomplete Cone Beam Data” to Tam.
[0048] In typical applications, a computer or other type of dedicated logic processor act as control logic processor for obtaining, processing, and storing image data is part of the CBCT system, along with one or more displays for viewing image results, as shown in
[0049] The subject matter of the present disclosure relates to digital image processing and computer vision technologies that process data from a digital image to recognize and thereby assign useful meaning to human-understandable objects, attributes, or conditions, and then to utilize the results obtained in further processing of the digital image.
[0050] Referring to the flowchart of
[0051] Continuing with the
[0052] Within each defined sub-volume from step S340 of
[0053] Using the mask or contour information obtained from step S360 of
[0054] The progression shown in
[0055] (i) define the overall 3D volume that includes the dentition;
[0056] (ii) define upper and lower sub-volumes within the overall volume;
[0057] (iii) generate 2D MIP images within each respective sub-volume;
[0058] (iv) delineate teeth within the MIP images to obtain 2D mask or contour information;
[0059] (v) apply the 2D mask or contour information to 3D segmentation.
[0060] Subsequent description gives more detail on each of the processing steps outlined in the
[0061] Plane Positioning Step S320
[0062]
[0063] Using conventional operator interface tools (not shown), the operator can perform various on-screen positioning tasks, including: [0064] (i) Rotate, zoom, and translate for both planes P1 and P2 individually or for the full composite image that includes the planes P1 and P2 and the displayed volume rendition. [0065] (ii) Specify the position of each individual plane P1, P2, such as using rotation and translation, with the volume maintained in a given position.
[0066]
[0067] An exemplary guideline is to provide plane placement that helps with the following: [0068] (i) Extracting the whole dentition section; [0069] (ii) Guiding the separation of upper and lower dentition sections (upper and lower jaws); [0070] (iii) Producing two MIP images each of which contains distinct teeth shapes that, in turn, facilitate generating satisfactory teeth 2D masks or teeth 2D contours that the subsequent automatic tooth segmentation process utilizes.
[0071] Approximate x, y, and z orthogonal coordinate axes are represented in each of
[0072] According to an alternate example embodiment of the present disclosure, planes P1 and P2 are automatically positioned by the system processor. To do this, system logic can execute a series of steps such as the following: [0073] (i) identify upper and lower tooth crowns in the CBCT volume; [0074] (ii) estimate an occlusal plane based on the identified crowns; and [0075] (iii) generate two planes approximately parallel to the estimated occlusal plane wherein the planes intersect their corresponding crown sections; wherein the two planes are apart from each other within a predetermined distance, such as within an exemplary distance of 15 pixels (assuming 300 microns per pixel).
[0076] VOI Extraction Step S330
[0077] The logic flow diagram of
[0078] Initially, the x, y, z axes shown in
[0079] Following the sequence of
[0080] A box computation step S333 computes z.sub.max and z.sub.min values of the extracting box according to the estimated z.sub.center value and the average tooth height, obtained from prior knowledge, such as stored values from statistical sampling or values entered for the particular patient. A computation step S336 then computes values x.sub.max, x.sub.min, y.sub.max, and y.sub.min that define the other two dimensions of the extracting box.
[0081] Jaw Segmentation Step S340
[0082] For orthodontic applications, the patient is asked to have the mouth closed during imaging, so that upper and lower teeth are in contact with each other. Therefore, jaw separation or segmentation is desirable for an automatic tooth segmentation system.
[0083]
[0084]
[0085]
[0086]
[0087]
[0088]
[0089] MIP Generation Step S350
[0090]
[0091] In an MIP formation step S352, processing forms two separate MIPs, one for maxilla sub-volume 50 and one for mandible sub-volume 52. Each respective MIP is formed using data content considered in the direction of a normal to the corresponding plane P1, P2. For each sub-volume 50, 52, this maximum intensity projection method begins, for example, at the intersection (with either plane P1 or P2) of a line that is parallel to the normal of the corresponding plane P1 or P2, assessing the intensity value of each voxel along the line, retaining the maximum (or higher) value voxel; continuing the same “assessing and retaining” projection method for all the voxel data along the same line, to the cusps of the respective teeth. This projection method repeats for all intersection points, resulting in an exemplary MIP image that contains more distinct patterns of the teeth compared to the surrounding background, as shown in
[0092] Continuing with the
[0093] Tooth Delineation Step S360
[0094]
[0095] Delineation uses a smoothed medial axis or center line C as a type of geometric spline 56 for the connected tooth region for each jaw sub-volume 50, 52. In
[0096]
[0097]
[0098] A Gaussian or other low-pass filter serves to smooth the length vector data and to reduce or eliminate spurious data and noise. The filtered length data are plotted as the oscillating bold curve in
[0099]
[0100]
[0101] Segmentation Step S370
[0102] Segmentation step S370 of the
[0103]
[0107] Cumulative segmentation results in all slices being grouped together to obtain 3D tooth segmentation results for the corresponding sub-volume.
[0108] According to an example embodiment of the present disclosure, after individual teeth are segmented separately (
[0109] Handling for False Positives/False Negatives
[0110] Identifying and compensating for false positives and false negatives can help to markedly improve the accuracy of segmentation step S370 of
[0113] An exemplary false negative 84 in tooth segmentation is presented in the axial view of
[0114] An example embodiment of the present disclosure addresses the task of reducing the number of false negatives of the type shown in
[0117] Concavity, particularly for exposed tooth surfaces, can often suggest a segmentation error with many types of teeth. According to an example embodiment of the present disclosure, the following method steps can be executed to differentiate a “correct” concave contour from an “erroneous” concave contour: [0118] 1. Comparison step. This step compares the segmentation result of the current slice with results from the previous, adjacent slice. This comparison can include identifying a region R1, as shown in
[0121] Another example of a false positive error related to ambiguous bone/root distinction is shown in
[0122] A method to correct for the exemplary false positive condition of
[0125] To prevent segmentation with leaking to outside of the tooth (or roots) region, additional forces can be introduced in the level-set energy functions as follows: [0126] (i) For region-driven level-set segmentation, as it tends to expand to neighboring teeth-of-non-interest or bones having similar intensity pixels, add a shrink-force to the level-set algorithm. This shrink-force can prevent “outside” leaking, in which false positive results spread and encroach upon true negative regions, that is, background regions (such as bones, teeth-of-non-interest). [0127] (ii) For edge-driven level-set segmentation, as it tends to snap to strong edges and to maintain them, an expand-force can be applied. Application of this force can help to prevent “inside” leaking, in which a false negative enters true positive regions, such as tooth or root regions.
[0128] Example embodiments of the present invention provide an automated tooth segmentation system that, beginning with a reconstructed 3D volume, identifies upper and lower jaw sub-volumes, generates and processes MIP image content for each sub-volume, and applies 2D MIP segmentation results to segmentation of the complete 3D volume image.
[0129] Consistent with an example embodiment of the present invention, a computer executes a program with stored software instructions that perform on image data accessed from an electronic memory, to provide panoramic presentation and tooth segmentation in accordance with the method described. As can be appreciated by those skilled in the image processing arts, a computer program of an example embodiment of the present invention can be utilized by a suitable, general-purpose computer system, such as a personal computer or workstation. However, many other types of computer systems can be used to execute the computer program of the present invention, including networked processors. The computer program for performing the methods of the present invention may be stored in a computer readable storage medium. This medium may comprise, for example; magnetic storage media such as a magnetic disk (such as a hard drive) or magnetic tape or other portable type of magnetic disk; optical storage media such as an optical disc, optical tape, or machine readable bar code; solid state electronic storage devices such as random access memory (RAM), or read only memory (ROM); or any other physical device or medium employed to store a computer program. The computer program for performing the methods of the present invention may also be stored on computer readable storage medium that is connected to the image processor by way of the internet or other communication medium. Those skilled in the art will readily recognize that the equivalent of such a computer program product may also be constructed in hardware.
[0130] It is noted that the computer program product of the present invention may make use of various image manipulation methods and processes that are well known. It will be further understood that the computer program product example embodiment of the present invention may embody methods and processes not specifically shown or described herein that are useful for implementation. Such methods and processes may include conventional utilities that are within the ordinary skill of the image processing arts. Additional aspects of such methods and systems, and hardware and/or software for producing and otherwise processing the images or co-operating with the computer program product of the present invention, are not specifically shown or described herein and may be selected from such methods, systems, hardware, components and elements known in the art.
[0131] The invention has been described in detail with particular reference to example embodiments, but it will be understood that variations and modifications can be affected that are within the scope of the invention. For example, the operator could enter equivalent bounding box information and seed information in any of a plurality of ways, including pointing to a particular tooth or other object using a touch screen or making a text entry on a keyboard, for example. The presently disclosed example embodiments are, therefore, considered in all respects to be illustrative and not restrictive. The scope of the present invention is indicated by the appended claims, and all changes that come within the meaning and range of equivalents thereof are intended to be embraced therein.