DETECTING AND MONITORING DEVELOPMENT OF A DENTAL CONDITION

20210045859 · 2021-02-18

Assignee

Inventors

Cpc classification

International classification

Abstract

A method, user interface and system for detecting and monitoring development of a dental condition. In particular, detecting and monitoring such a development by comparing digital 3D representations of the patient's set of teeth recorded at a first and a second point in time. For example, determining tooth movement for at least one tooth between the first and second point in time based on derived distances.

Claims

1. A method for determining an orientation and position of a dental implant arranged in a patient's jawbone, the method comprising obtaining a digital 3D representation of a healing abutment that is arranged in the dental implant; aligning a CAD model of the healing abutment with the obtained digital 3D representation of the healing abutment; determining an orientation and position of the obtained digital 3D representation of the healing abutment based on the alignment of the CAD model and obtained digital 3D representation; and deriving an orientation and position of the dental implant in the patient's jaw based on the determined orientation and position of the obtained digital 3D representation.

2. The method according to claim 1, wherein deriving the orientation and position of the dental implant in the patient's jaw is based on a code providing information relating to the healing abutment.

3. The method according to claim 2, wherein the information comprises shape and size of the healing abutment.

4. The method according to claim 1, wherein deriving the orientation and position of the dental implant is performed during osseointegration of the dental implant in the patient's jaw.

5. The method according to claim 1, wherein the healing abutment, corresponding to the digital 3D representation, arranged in the dental implant comprises the healing abutment being firmly attached to the dental implant such that a change in the dental implant position and orientation results in a corresponding change in the position and orientation of the healing abutment.

6. The method according to claim 1, further comprising deriving orientation and position of the dental implant in the patient's jaw for digital 3D representations of the healing abutment at different times during the osseointegration.

7. The method according to claim 6, further comprising detecting change in orientation and position of the dental implant during the osseointegration based on comparing orientation and position of the dental implant derived for the different times.

8. The method according to claim 7, further comprising raising an alarm when the detected change in the orientation and position of the dental implant exceeds a value during the osseointegration.

9. The method according to claim 7, further comprising designing an abutment and crown in response to the change in the dental implant position and orientation during the osseointegration.

10. A computer program product comprising program code means for causing a data processing system to perform method for determining an orientation and position of a dental implant arranged in a patient's jawbone, the method comprising obtaining a digital 3D representation of a healing abutment that is arranged in the dental implant; aligning a CAD model of the healing abutment with the obtained digital 3D representation of the healing abutment; determining an orientation and position of the obtained digital 3D representation of the healing abutment based on the alignment of the CAD model and obtained digital 3D representation; and deriving an orientation and position of the dental implant in the patient's jaw based on the determined orientation and position of the obtained digital 3D representation.

11. The computer program product according to claim 10, wherein deriving the orientation and position of the dental implant in the patient's jaw is based on a code providing information relating to the healing abutment.

12. The computer program product according to claim 11, wherein the information comprises shape and size of the healing abutment.

13. The computer program product according to claim 10, wherein deriving the orientation and position of the dental implant is performed during osseointegration of the dental implant in the patient's jaw.

14. The computer program product according to claim 10, wherein the healing abutment, corresponding to the digital 3D representation, arranged in the dental implant comprises the healing abutment being firmly attached to the dental implant such that a change in the dental implant position and orientation results in a corresponding change in the position and orientation of the healing abutment.

15. The computer program product according to claim 10, wherein the method further comprises deriving orientation and position of the dental implant in the patient's jaw for digital 3D representations of the healing abutment at different times during the osseointegration.

16. The computer program product according to claim 15, wherein the method further comprises detecting change in orientation and position of the dental implant during the osseointegration based on comparing orientation and position of the dental implant derived for the different times.

17. The computer program product according to claim 16, wherein the method further comprises raising an alarm when the detected change in the orientation and position of the dental implant exceeds a value during the osseointegration.

18. The computer program product according to claim 16, wherein the method further comprises designing an abutment and crown in response to the change in the dental implant position and orientation during the osseointegration.

Description

BRIEF DESCRIPTION OF THE DRAWINGS

[0124] The above and/or additional objects, features and advantages of the present invention, will be further elucidated by the following illustrative and non-limiting detailed description of embodiments of the present invention, with reference to the appended drawings, wherein:

[0125] FIG. 1 shows a workflow for an embodiment.

[0126] FIG. 2A-2B show a set of teeth and segmentation of a tooth.

[0127] FIGS. 3A-3C illustrate an embodiment for detecting gingival retraction.

[0128] FIGS. 4A-4D illustrate how anatomically correct measurement of tooth movement can be made.

[0129] FIG. 5 shows an outline of a system.

DETAILED DESCRIPTION

[0130] In the following description, reference is made to the accompanying figures, which show by way of illustration how the invention may be practiced.

[0131] FIG. 1 shows a workflow for an embodiment of the method for detecting development of a dental condition for a patient's set of teeth between a first and a second point in time. The workflow 100 includes steps 102, 103 for obtaining a first and a second digital 3D representation of the patient's set of teeth. The digital 3D representations can be recorded using an intra oral scanner, such as the TRIOS 3 intra oral scanner by 3shape A/S which can record both the topography and color of the patient's set of teeth. The recorded digital 3D representations then expresses both the geometry and colors of the scanned teeth at the first and second points in time. Color calibrating the scanner regularly or just prior to the scanning provides that the measured colors are true and that the color recorded at one visit at the dentist can be compared with the colors measured at another visit.

[0132] In step 104 the first and second digital 3D representations are globally aligned using e.g. a 3-point alignment where 3 corresponding regions are marked on the first and second digital 3D representations. The aligned digital 3D representations can then be visualized in the same user interface and comparisons between shapes and sizes of teeth can be made straightforward. The global alignment of the digital 3D representations can be performed using a computer implemented algorithm, such as an Iterative Closest Point (ICP) algorithm, employed to minimize the difference between digital 3D representations.

[0133] In step 105 the aligned first and second digital 3D representations are compared e.g. by calculating a difference map showing the distance between the digital 3D representations at the different parts of the teeth of teeth. Such as difference map can e.g. be used for monitoring tooth movement during an orthodontic treatment. Based on the comparison a change in a parameter relating to the dental condition can be detected in step 106 and the change in the parameter can be correlated with a development of a dental condition in step 107.

[0134] When the dental condition corresponds to caries and the development of the caries is monitored using change in tooth color from white to brown in the infected region, the global alignment and comparison of the digital 3D representations provide that a change in the tooth color to a more brownish color in a region of the teeth can be detected and the region can be visualized to the operator. The change in the color can be measured using color values of e.g. the RGB system and can be correlated with knowledge of the usual changes in tooth colors during development of caries.

[0135] FIGS. 2A-2B show a digital 3D representation of a set of teeth and segmentation of the digital 3D representation to create a 3D model of a tooth. The digital 3D representation 230 has topography data for four anterior teeth 2311, 2312, 2313, 2314 and for a portion of the corresponding gingiva with the gingival boundary 232 as indicated in FIG. 2A. The segmentation of the digital 3D representation provides a 3D tooth model 233 which has the shape of the corresponding tooth part of the digital 3D representation 2312 and is bounded by the gingival boundary 232. In FIG. 2B the 3D tooth model 233 is still arranged along with the other parts of the digital 3D representation according to the arrangement of the tooth in the digital 3D representation.

[0136] FIGS. 3A-3C illustrate an embodiment for detecting gingival retraction at the patient's left central incisor 3313. This tooth is segmented in both the first 340 and second 341 digital 3D representation of the set of teeth showing the two central incisors in the lower jaw and the gingival boundary 332 as seen inn FIGS. 3A and 3B. The change in the position of the gingival boundary is so small that when the two digital 3D representations are seen separately the change is hardly visible. A section having topography data relating to both the tooth surface and the gingiva is selected and the two sections are locally aligned based on the tooth topography data. The local alignment can be performed using iterative closest point algorithm. All data of the sections are aligned according to this local transformation as seen in FIG. 3C and the gingival retraction from the boundary 3321 in the first digital 3D representation to the 3322 boundary in the second digital 3D representation becomes clearly visible. The gingival retraction can now be measured as the distance between the two boundaries.

[0137] FIG. 4A shows cross sections of a first 3D tooth model 451 and a second 3D tooth model 451 segmented from a first and a second digital 3D representation, respectively. The first digital 3D representation can for example represent the patient's teeth at the onset of an orthodontic treatment and the second digital 3D representation at some point during the treatment. A transformation T which aligns the first 3D tooth model with the second 3D tooth model is determined and applied to the first 3D tooth model to provide that the two tooth models are locally aligned as illustrated in FIG. 4B. In FIG. 4C three portions 4551, 4561 and 4571 are selected on the first 3D tooth model 451. Since the first and second 3D tooth models are locally aligned the anatomically corresponding portions 4552, 4562 and 4572 can readily and precisely be identified on the second 3D tooth model 451. In FIG. 4D portions 4551, 4561 and 4571 are marked on corresponding portions of the first digital 3D representation 460 and portions 4552, 4562 and 4572 are marked on the corresponding portions of the second digital 3D representation 461. The first and second digital 3D representations are global aligned based e.g. on the other teeth of the same quadrant or same arch. The anatomical correct distance between the marked regions can then be determined and based on these distances a measure for the movement of the tooth between the first and second point in time can be derived.

[0138] In short the workflow described here has the following steps: [0139] selecting one or more corresponding regions on the locally aligned segmented teeth, [0140] globally aligning the first and second digital 3D representations; [0141] identifying the selected corresponding regions on the globally aligned first and second digital 3D representations [0142] deriving the distances between the selected corresponding regions on the globally aligned first and second digital 3D representations [0143] determining the tooth movement based on the derived distances.

[0144] In a computer program product configured for implementing the method the portions on the first 3D tooth model can be selected by an operator or by the computer program product when this is configured for detecting appropriate portions, such as characteristic portions on the cusp. The selected portion can also be the entire tooth surface such that a distance map is derived showing the movement for the entire surface.

[0145] Other workflows can also be used to measure the distance such as: [0146] selecting one or more corresponding regions on the locally aligned segmented teeth, [0147] arranging the first and second 3D tooth models according to the global alignment; [0148] deriving the distances between the selected corresponding regions in the global alignment of the first and second 3D tooth models; [0149] determining the tooth movement based on the derived distances.

[0150] FIG. 5 shows a schematic of a system according to an embodiment. The system 570 has a computer device 571 with a data processing unit in the form of a microprocessor 572 and a non-transitory computer readable medium 573 encoded with a computer program product providing a digital tool for determining the movement of teeth e.g. during an orthodontic treatment. The system further has a visual display unit 576, a computer keyboard 574 and a computer mouse 575 for entering data and activating virtual buttons of a user interface visualized on the visual display unit 576. The visual display unit 576 can e.g. be a computer screen. The computer device 571 is capable of storing obtained digital 3D representations of the patient's teeth in the computer readable medium 573 and loading these into the microprocessor 572 for processing. The digital 3D representations can be obtained from a 3D color scanner 577, such as the 3Shape TRIOS 3 intra-oral scanner, which is capable of recording a digital 3D representation containing both geometrical data and color data for the teeth.

[0151] Besides color and geometry data the digital 3D representation can also include diagnostic data, such fluorescence data obtained using an intra-oral scanner.

[0152] The computer readable medium 573 can further store computer implemented algorithms for segmenting a digital 3D representation to create digital 3D models of the individual teeth and for selecting regions on the surface for a local alignment. When digital 3D models for the same tooth is created from different digital 3D representations, such as digital 3D representations recorded at different points in time, the digital 3D models can be locally aligned using e.g. Iterative Closest Point algorithms (ICP) for minimizing the distance between the surfaces of the digital 3D representations. The digital 3D representations of the patient's entire set of teeth or sections thereof can be globally aligned also using such ICP algorithms. When the digital 3D representations of the teeth are globally aligned with the anatomically correct corresponding regions of a given tooth identified by the local alignment procedure applied to the digital 3D model of that tooth, the precise measure of the movement of the tooth between the points in time where the two digital 3D representations were recorded can be determined.

[0153] When the tooth movement has been determined it can be visualized to the operator in the visual display unit 576 e.g. as a distance map or using a cross sectional view of the 3D tooth models or the digital 3D representations.

[0154] The digital 3D models of the individual teeth can be stored on the computer readable medium and be re-used at the next visit for the identification of individual teeth in a digital 3D representation recorded at the next visit.

[0155] Although some embodiments have been described and shown in detail, the invention is not restricted to them, but may also be embodied in other ways within the scope of the subject matter defined in the following claims. In particular, it is to be understood that other embodiments may be utilized and structural and functional modifications may be made without departing from the scope of the present invention.

[0156] In device claims enumerating several means, several of these means can be embodied by one and the same item of hardware. The mere fact that certain measures are recited in mutually different dependent claims or described in different embodiments does not indicate that a combination of these measures cannot be used to advantage.

[0157] A claim may refer to any of the preceding claims, and any is understood to mean any one or more of the preceding claims.

[0158] It should be emphasized that the term comprises/comprising when used in this specification is taken to specify the presence of stated features, integers, steps or components but does not preclude the presence or addition of one or more other features, integers, steps, components or groups thereof.

[0159] The features of the method described above and in the following may be implemented in software and carried out on a data processing system or other processing means caused by the execution of computer-executable instructions. The instructions may be program code means loaded in a memory, such as a RAM, from a storage medium or from another computer via a computer network. Alternatively, the described features may be implemented by hardwired circuitry instead of software or in combination with software.