Machine learning device, estimation device, non-transitory computer readable medium, and learned model
11631177 · 2023-04-18
Assignee
Inventors
- Megumi Nakao (Kyoto, JP)
- Toyofumi Yoshikawa (Kyoto, JP)
- Junko Tokuno (Kyoto, JP)
- Hiroshi Date (Kyoto, JP)
- Tetsuya Matsuda (Kyoto, JP)
Cpc classification
G16H50/20
PHYSICS
G06T17/20
PHYSICS
A61B2034/105
HUMAN NECESSITIES
A61B34/10
HUMAN NECESSITIES
International classification
A61B34/10
HUMAN NECESSITIES
G06T17/20
PHYSICS
G06T7/246
PHYSICS
Abstract
A machine learning device includes: a generation unit generating a first shape model representing a shape of an object before deformation and a second shape model representing a shape of the object after the deformation based on measurement data before and after the deformation; and a learning unit learning a feature amount including a difference value between each micro region and another micro region that constitute the first shape model, and a relation providing a displacement from the each micro region of the first shape model to each corresponding micro region of the second shape model.
Claims
1. A machine learning device comprising: a generation unit generating a first shape model representing a shape of an object before deformation and a second shape model representing a shape of the object after the deformation based on measurement data before and after the deformation; and a learning unit learning a feature amount including a difference value between each micro region and another micro region that constitute the first shape model, and a relation providing a displacement from the each micro region of the first shape model to each corresponding micro region of the second shape model.
2. The machine learning device according to claim 1, wherein, when the generation unit generates a third shape model with a change in positions of vertexes, which are elements of the first shape model, and a fourth shape model with a change in positions of vertexes, which are elements of the second shape model, the learning unit also learns a feature amount including a difference value between each micro region and another micro region that constitute the third shape model, and a relation providing a displacement from the each micro region of the third shape model to each corresponding micro region of the fourth shape model.
3. The machine learning device according to claim 1, wherein the feature amount is defined by at least one of coordinates giving each micro region, a gradient or a normal vector defining each micro region, a curvature defining each micro region, Voronoi area of each micro region, a shrinkage ratio of each micro region, and a deviation amount between each micro region and a corresponding region of an average shape model.
4. The machine learning device according to claim 1, wherein the difference value giving the feature amount is calculated with the micro regions of at least 5% of total micro regions constituting the first shape model.
5. The machine learning device according to claim 1, wherein the learning unit learns the relation by using a kernel regression model.
6. The machine learning device according to claim 1, wherein the measurement data before and after the deformation is medical image data.
7. The machine learning device according to claim 6, wherein the deformation is deaerated deformation of a lung, medical image data of a lung of a patient having treated pneumothorax is used as the measurement data before the deformation due to deaeration, and medical image data of the lung of the patient before treatment of pneumothorax is used as the measurement data after the deformation due to deaeration.
8. The machine learning device according to claim 6, wherein the deformation includes deformation of an organ or movement of a position of an organ due to breathing, time variation, or differences in specimens.
9. The machine learning device according to claim 1, wherein the first shape model and the second shape model are three-dimensional vertex models.
10. The machine learning device according to claim 1, wherein the first shape model and the second shape model are three-dimensional vertex models, and the difference value between each micro region and another micro region is a difference between vertexes.
11. A non-transitory computer readable medium storing a program causing a computer to execute: a function of generating a first shape model representing a shape of an object before deformation and a second shape model representing a shape of the object after the deformation based on measurement data before and after the deformation; and a function of learning a feature amount including a difference value between each micro region and another micro region of the first shape model, and a relation providing a displacement from the each micro region of the first shape model to each corresponding micro region of the second shape model.
12. An estimation device comprising: an estimation unit, when any measurement data is given, using a learned model having learned a feature amount including a difference value between each micro region and another micro region in a first shape model generated from measurement data before deformation, and a relation providing a displacement from the each micro region of the first shape model to each corresponding micro region of a second shape model after the deformation, to thereby estimate a shape after the deformation corresponding to the any measurement data.
13. The estimation device according to claim 12, wherein the estimation unit calculates a feature amount including a difference value between a micro region and another micro region of a shape model generated for the inputted measurement data, and estimates a displacement of a feature amount corresponding to each micro region by interpolating a relation stored in the learned model.
14. A non-transitory computer readable medium storing a program causing a computer to execute: a function of, when any measurement data is given, using a learned model having learned a feature amount including a difference value between each micro region and another micro region in a first shape model generated from measurement data before deformation, and a relation providing a displacement from the each micro region of the first shape model to each corresponding micro region of a second shape model after the deformation, to thereby estimate a shape after the deformation corresponding to the any measurement data.
15. A non-transitory computer readable medium storing a program causing a computer to execute a function comprising: providing a relation having a feature amount including a difference value between each micro region and another micro region in a first shape model generated from measurement data before deformation as an input, and a displacement from the each micro region of the first shape model to each corresponding micro region of a second shape model after the deformation as an output, and, when any measurement data is given, computing a displacement corresponding to each micro region of a shape model corresponding to the any measurement data by using the relation, to thereby estimate a shape after the deformation corresponding to the any measurement data.
16. A machine learning device comprising: a generation unit generating shape models representing individual shapes of a first object and an average shape model based on a plurality of measurement data items of the first object; and a learning unit learning a distribution of a feature amount including a difference value between each micro region constituting the shape model representing the individual shape and each corresponding micro region of the average shape model, and a relation between a distribution of the difference value and a region where a second object having a constraint relation with the first object exists.
17. The machine learning device according to claim 16, wherein the constraint relation includes a relation in which the second object is adjacent to or connected to the first object.
18. A non-transitory computer readable medium storing a program causing a computer to execute: a function of generating shape models representing individual shapes of a first object and an average shape model based on a plurality of measurement data items of the first object; and a function of learning a distribution of a feature amount including a difference value between each micro region constituting the shape model representing the individual shape and each corresponding micro region of the average shape model, and a relation between a distribution of the difference value and a region where a second object having a constraint relation with the first object exists.
Description
BRIEF DESCRIPTION OF DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
(21)
(22)
(23)
(24)
(25)
(26)
(27)
(28)
(29)
(30)
(31)
(32)
(33)
(34)
(35)
(36)
(37)
(38)
(39)
(40)
(41)
DESCRIPTION OF EMBODIMENTS
(42) Hereinafter, with reference to attached drawings, exemplary embodiments according to the present invention will be described in detail.
Exemplary Embodiment 1
(43) <System Configuration>
(44)
(45) The estimation system 1 shown in
(46) In the case of the exemplary embodiment, input of the measurement data to the machine learning device 10 may be performed directly from a database (DB) 30 storing the measurement data, or may be performed via a network 40. The network 40 here may be a local area network or the Internet.
(47) In addition, the generated learned models and the parameters (update parameters) updating the learned models are provided to the deformation estimation device 20 directly or via the network 40. However, the learned models and update parameters may be provided to the deformation estimation device 20 via a not-shown storage medium. The update parameter is the value of the coefficient of the deformation function updated by the learning, for example.
(48) The deformation estimation device 20 inputs the measurement data to be estimated directly from a terminal 50 or via the network 40, and outputs the result of estimation (estimation result) to the terminal 50 directly or via the network 40. The terminal 50 here may be, for example, a surgery support system or a radiotherapy device.
(49) <Machine Learning Device>
(50)
(51) The machine learning device 10 shown in
(52) A storage device 14, for example, is configured with a hard disk device, and stores application programs learning the relation between before and after the deformation and the learned models that are the results of the learning. Note that the storage device 14 is not limited to the internal device, but may be an external device, or a networked storage.
(53) An input device 15, for example, is a keyboard or a mouse, and is used for operation inputting. An output device 16 is a display device or a printing device used to display operation screens or to print information. Note that, in the case where the machine learning device 10 is implemented as a server, the input device 15 and the output device 16 may be provided on the not-shown terminal side.
(54) A communication device 17 is used to communicate with external devices. The communication device 17 is used to input measurement data from the database 30 and to output the learned models or the update parameters to the deformation estimation device 20 (including distribution).
(55) The CPU 11 and each component are connected via a bus 18 and not-shown signal wires.
(56) Note that the machine learning device 10 can also be implemented as dedicated hardware.
(57)
(58) In
(59) The machine learning device 10 in the exemplary embodiment functions as a shape model generation unit 101 generating vertex models from the pre-deformation image I.sub.(i)I and the post-deformation image I.sub.(i)D, a model upsampling unit 102 generating multiple vertex models with different vertex dispositions from the generated vertex model, and a deformation learning unit 103 learning the relation between the feature amount of each micro region of the vertex model generated by upsampling and the displacement between the micro regions before and after the deformation.
(60) All the vertex models in the exemplary embodiment are three-dimensional vertex models. Note that the three-dimensional vertex model may be defined by vertexes of a polygon mesh representing a surface by polygons, or by vertexes of a voxel mesh representing both a surface and an interior by voxels.
(61) Unlike general learning, the learning here is performed, not on the entire vertex model, but per micro region unit constituting the vertex model. The deformation function f, which is the result of the learning, is output as a learned model.
(62) There is no particular restriction on the object to be learned, but in the exemplary embodiment, the object is assumed to change the shape or position (or region) thereof. In the exemplary embodiment, the target of the learning is the lung, liver, stomach, bowel, kidney, pancreas and other organs of the specimen (for example, human or animal).
(63)
(64) The laboratory beagle dog was used to generate the displacement models A, B and C. In the experiment, the bronchial pressure was changed from 14 cmH.sub.2O to 2 cmH.sub.2O, and the shape of the lung was measured with a three-dimensional CT.
(65) In the exemplary embodiment, medical image data items captured the specimen's organs are used as the pre-deformation image I.sub.(i)I and the post-deformation image I.sub.(i)D. Examples of the medical image data include the CT image data, the magnetic resonance imaging (MRI) image data, the X-ray image data, the ultrasonic wave image data, and the positron emission tomography (PET) image data. The medical image data may be the two-dimensional image data or the three-dimensional image data. The two-dimensional image data is represented by pixels defining the two-dimensional space, and the three-dimensional image data is represented by voxels (cubic unit cells), polygons, and the like defining the three-dimensional space. The medical image data is an example of the measurement data.
(66) In the exemplary embodiment, the pre-deformation image I.sub.(i)I and the post-deformation image I.sub.(i)D are assumed to be the three-dimensional image. The three-dimensional image is defined, for example, by coordinates and voxel values. The voxel value differs depending on the image-capturing technique.
(67) Let us return to the description of
(68) The shape model generation unit 101 generates the vertex models from the pre-deformation image I.sub.(i)I and the post-deformation image I.sub.(i)D.
(69)
(70) The shape model generation unit 101 shown in
(71) The sampling unit 101A, for example, samples each of the pre-deformation image I.sub.(i)I and the post-deformation image I.sub.(i)D to generate shape data of the voxel mesh structure or the polygon mesh structure. Here, it is assumed that the shape data corresponding to the pre-deformation image I.sub.(i)I is S.sub.(i)I and the shape data corresponding to the post-deformation image I.sub.(i)D is S.sub.(i)D.
(72) The size of the pre-deformation image I.sub.(i)I and the post-deformation image I.sub.(i)D differs depending on the differences in the specimens and the image capturing conditions. In addition, there are limitations to unify the position and physical condition of the specimens during image capturing. For this reason, the number of vertexes of the voxels included in the independently-formed shape data S.sub.(i)I and S.sub.(i)D is inconsistent in general.
(73) In the case of
(74)
(75) In the case of
(76) Let us return to the description of
(77) When the shape data items S.sub.(i)I and S.sub.(i)D are obtained, the alignment unit 101B performs alignment between the shape data items. Specifically, the process of aligning the number of vertexes in each shape data is performed.
(78) In
(79) The pre-deformation shape model M.sub.(i)I is an example of the first shape model, and the post-deformation shape model M.sub.(i)D is an example of the second shape model.
(80) In
(81) Generation of the shape model M.sub.(i) with the unified number of vertexes enables statistical learning of deformation of the object.
(82) Note that, in the case where the shape data items S.sub.(i)I and S.sub.(i)D are inputted to the shape model generation unit 101 (refer to
(83)
(84) The alignment by the alignment unit 101B (refer to
(85)
(86) The two-stage alignment shown in
(87) Let us return to the description of
(88) The model upsampling unit 102 upsamples each of the shape models M(i)I and M(i)D generated by the shape model generation unit 101 to generate multiple shape models with similar shapes and structures but with different vertex locations.
(89) However, the model upsampling section 102 is not indispensable in the case of dealing with objects that are easy to acquire images before and after deformation.
(90) On the other hand, in the case of dealing with objects that are not easy to acquire images before and after transformation (for example, to deal with medical image data), it is possible to increase the number of teacher data items used for learning by providing the model upsampling unit 102. Even in the case of dealing with objects that are easy to acquire images, the number of the shape models M.sub.(i)I and M.sub.(i)D used for teacher data can be increased from fewer images by using the model upsampling unit 102.
(91)
(92) In the case of
(93) Here, the pre-deformation shape models M.sub.(i1)I, . . . , M.sub.(ik)I are an example of the third shape model, and the post-deformation shape models M.sub.(i1)D, . . . , M.sub.(ik)D are an example of the fourth shape model.
(94) Of course, the generated shape models M.sub.(i1)I, . . . , M.sub.(ik)I and M.sub.(i1)D, . . . , M.sub.(ik)D are all composed of 500 vertexes.
(95) In this way, upsampling generates the shape models with different vertex locations because, in the exemplary embodiment, the displacements are learned in the small region units of the shape model (for example, vertex units).
(96) As described above, the upsampling operations by the model upsampling unit 102 can be omitted in the case where a sufficient number of measurement data items are available in advance.
(97)
(98)
(99) The method shown in
(100) As a result, each vertex (indicated by a cross in the figure) that was on the surface before rotation is moved to the position indicated by a black circle. In the rotation and linear movement, each vertex in the shape model is repositioned so that relative positional relationship with the vertex on the surface is preserved. For example, in the case where the entire shape model is represented with tetrahedral elements, it is possible to parameterize the relative position of the interior of the tetrahedron.
(101) Note that, if the center of rotation is changed, different upsampling models can be generated even if the same processing process is performed. Similarly, the different upsampling models can be generated by changing the direction of rotation. The different upsampling models can also be generated by changes in the rules during linear movement.
(102) The upsampling shown in
(103)
(104)
(105) The method shown in
(106) As a result, each vertex (indicated by a cross in the figure) that was on the surface before enlarging conversion is moved to the position indicated by a black circle. In the enlarging conversion and linear movement, each vertex in the shape model is repositioned so that relative positional relationship with the vertex on the surface is preserved.
(107) Note that, if the position of the center used for enlarging conversion is changed, different upsampling models can be generated even if the same processing process is performed. Similarly, the different upsampling models can be generated by changing the enlargement factor. The different upsampling models can also be generated by changes in the rules during linear movement.
(108) The upsampling shown in
(109) Note that, as the method of perturbation (reposition) of the vertex position by upsampling, the method of repositioning each vertex to approximately match the original shape after the reducing conversion, or randomly perturbating the vertex position may be used.
(110) Moreover, in the above description, the vertexes on the surface among the rotated vertexes are aligned to move linearly to the surface of the original shape model; however, the alignment by linear movement may be performed under a weak constraint. The weak constraint means that errors within a predetermined range are allowed.
(111) The upsampling operation by the model upsampling unit 102 can generate multiple shape models with similarly represented shapes and structures but with different vertex locations. In other words, teacher data can be increased.
(112) Let us return to the description of
(113) The deformation learning unit 103 learns the relation between the pre-deformation shape models M.sub.(i1)I to M.sub.(ik)I generated by the above-described process and the post-deformation shape models M.sub.(i1)D to M.sub.(ik)D.
(114) In the case of the exemplary embodiment, the deformation is learned per micro region (for example, vertex) unit.
(115)
(116)
(117)
(118) In the exemplary embodiment, the feature amount data x.sub.i regarding each vertex i of the pre-deformation shape model is used as the input, and the displacement amount data y.sub.i regarding the corresponding vertex i before and after deformation is used as the output. At the learning stage, both the feature amount data x.sub.i and the displacement amount data y.sub.i are given, and the deformation function f defining the relation between the two values is learned.
(119) In the exemplary embodiment, the difference between micro regions on the same shape model (for example, the difference between vertexes d) is used as the feature amount data x.sub.i.
(120)
(121) In actuality, the difference d is calculated between one vertex i and 500 vertexes (micro regions). Here, when the coordinates of each vertex are expressed by v.sub.i and v.sub.j, the difference between vertexes d.sub.ij with respect to vertex i is expressed as d.sub.ij=(where j=1, 2, 3, . . . , 500).
(122) Consequently, the feature amount data x.sub.i with respect to the vertex i is defined as a set of 500 differences between vertexes d.sub.ij. As will be described later, the difference between vertexes d.sub.ij is an example of the feature amount data x.sub.i.
(123) In
(124)
(125) The deformation learning unit 103 shown in
(126) The feature amount calculation unit 103A in the exemplary embodiment calculates the feature amount data x.sub.i per each vertex of the shape models M.sub.(i1)I to M.sub.(ik)I generated from the measurement data before deformation given for learning.
(127) As described above, in the case of the exemplary embodiment, the feature amount data x.sub.i is given as the difference between vertexes d.sub.ij for all vertexes constituting the shape model M.sub.(i)I generated from the measurement data before deformation inputted for learning. In other words, 500 feature amount data items x.sub.i are calculated for a single shape model M.sub.(i)I.
(128) In the case of
(129) The displacement calculation unit 103B in the exemplary embodiment calculates the displacement amount data y.sub.i between the corresponding vertexes of the shape models in the relation of before and after the deformation. For a single shape model before and after deformation, 500 displacement amount data items y.sub.i are calculated.
(130) The machine learning unit 103C in the exemplary embodiment learns the deformation function f(i) expressing the relation between the feature amount data x.sub.i for each vertex i of the pre-deformation shape model M(j)I and the displacement amount data y.sub.i between the shape models before and after deformation for the same vertex i.
(131)
(132) In
(133) In the exemplary embodiment, the kernel regression model represented by expression 1 is used as a regression model representing the deformation function f.
(134)
(135) Here, k(x.sub.i, x) is the kernel function given by the following expression.
K(x.sub.i,x)=exp(−β∥x.sub.i−x∥.sup.2)
Note that ∥x.sub.i−x∥.sup.2 is the L.sup.2 norm corresponding to the difference between vertexes d.sub.ij between the vertex i and the other vertex j on the same shape model. In other words, k(x.sub.i, x) corresponds to the feature amount data x.sub.i for the vertex i. The individual a.sub.ik (x.sub.i, x) corresponds to f.sub.i (i=1, 2, 3) in
(136) The displacement calculation unit 103B in the exemplary embodiment learns the coefficient a.sub.i for each vertex i. The coefficient a.sub.i is an update parameter.
(137) The set of learned coefficients a.sub.i may be provided to the deformation estimation device 20 (refer to
(138) <Specific Examples of Learned Models>
(139) For example, medical image data of human lungs can be obtained as pre-surgical and post-surgical CT data of a patient with pneumothorax.
(140) The pre-surgical CT data is CT data of collapsed lung and the post-surgical CT data is CT data of normal lung (after treatment).
(141) In this case, the post-surgical CT data is used as the pre-deformation medical image data and the pre-surgical CT data is used as the post-deformation medical image data to learn the relation before and after the deformation as the learned model. The learned model can be used to predict lung deformation during surgery in the case where a patient with ordinary lung has surgery.
(142) <Deformation Estimation Device>
(143)
(144) The deformation estimation device 20 shown in
(145) A storage device 24, for example, is configured with a hard disk device, and stores application programs estimating the deformed shape of any measurement data given by the terminal 50 (refer to
(146) An input device 25, for example, is a keyboard or a mouse, and is used for operation inputting. An output device 26 is a display device or a printing device used to display operation screens or to print information. Note that, in the case where the deformation estimation device 20 is implemented as a server, the input device 25 and the output device 26 may be provided on the not-shown terminal side.
(147) A communication device 27 is used to communicate with external devices. The communication device 27 is used to input the measurement data from the terminal 50 and to output the estimation results to the terminal 50.
(148) The CPU 21 and each component are connected via a bus 28 and not-shown signal wires.
(149) Note that the deformation estimation device 20 can also be implemented as dedicated hardware.
(150)
(151) The deformation estimation device 20 in
(152)
(153) When any measurement data M.sub.(i)I is given from terminal 50 (refer to
(154) In
(155) Further, the deformation estimation unit 202 generates the deformation function f.sub.i acquired by linearly interpolating the learned models (namely, the deformation functions f.sub.2 and f.sub.4) regarding the multiple extracted shape models M.sub.(i2)I and M.sub.(i4)I, and estimates the displacement amount data y.sub.i corresponding to the feature amount data x.sub.i by using the generated deformation function f.sub.i.
(156) In the exemplary embodiment, the deformation function f.sub.i is generated by use of linear interpolation, but nonlinear interpolation can be used to generate the deformation function f.sub.i.
(157)
(158) The output screen 300 has a pre-deformation observation image section 301 disposed on the left and an estimation image section 302 corresponding to the post-deformation shape disposed on the right.
(159) In the case of
(160) The lung shown in the estimation image section 302 is an estimated image during surgery. The lung during surgery is deflated by deaerated deformation. As shown in
(161) <Estimation Error>
(162) Hereinafter, description will be given of the error between the estimation results by the above technique and the actually measured values.
Experimental Conditions
(163) Measurement of living lung data of 11 beagle dogs
(164) Modeling of the superior lobe with 204 vertexes and the inferior lobe with 200 vertexes
(165) Upsampling of the living lung data measured at the aerated condition by enlarging and reducing the data to ±2.5 mm, ±5.0 mm, ±7.5 mm
(166) Use of 80% of the measurement data for training data and 20% for test data
Evaluation
(167) Comparison of the following regression models using the Hausdorff distance as the error index
(168) Regression models (multiple linear regression models, Lasso regression models, kernel regression models)
(169) Feature amount data (4 types)
(170) The Hausdorff distance is the distance given as the maximum value of deviation between two shapes.
(171)
(172) The values shown in
(173) In the case of
(174) The kernel regression model used in the exemplary embodiment was approximately one-third of the multiple linear regression model or the Lasso regression model (L1 regularization).
(175)
(176) The four types of feature amount data are assumed to be the following feature amount groups 1 to 4 regarding the vertex i.
(177) Feature amount group 1: Coordinates v.sub.i, normal vector n.sub.i, curvature (second derivative, discrete Laplacian) l.sub.i, difference between vertexes d.sub.ij, shrinkage ratio s.sub.i, Voronoi area r.sub.i
(178) Feature amount group 2: Difference between vertexes d.sub.ij
(179) Feature amount group 3: Difference between vertexes d.sub.ij, normal vector n.sub.i×100
(180) Feature amount group 4: Coordinates v.sub.i, normal vector n.sub.i, curvature (second derivative, discrete Laplacian) l.sub.i, shrinkage ratio s.sub.i, Voronoi area r.sub.i
(181) The feature amount group 2 corresponds to the feature amount data x.sub.i; used in the previous description.
(182) From the diagram shown in
(183) From the above, it is found that inclusion of the difference between vertexes d.sub.ij in the feature amount data is useful for increasing the estimation accuracy.
(184)
(185)
(186) The specimens in
(187) Only the specimen 2 has the estimation error of more than 5 mm in the superior lobe, and the specimens 2 and 8 have the estimation error of more than 5 mm in the inferior lobe.
(188) In this experiment, the estimation error is relatively large in the lung with a much larger inferior lobe than the superior lobe, but in the lung with the superior lobe smaller than the inferior lobe and the superior lobe approximately the same size as the inferior lobe, the estimation error is small.
(189) Note that, for the models estimated by the statistical mean displacement for the entire shape, errors of 10 mm or more were found; accordingly, reduction in the estimation error can be found in the technique of the exemplary embodiment.
(190)
(191)
(192) The ratio of the horizontal axis represents the case in which all the vertexes constituting the shape model M.sub.(i)I are used to calculate the difference between vertexes d.sub.ij as 100%. Consequently, in the case where the number of vertexes is 500, the ratio of the difference between vertexes constituting the feature amount data x.sub.i being 10% means that the feature amount data x.sub.i for the vertex i is given by the 50 differences between vertexes d.sub.ij. The vertical axis is in the unit of mm.
(193) In the above description, the example in which the feature quantity data x.sub.i is given as the difference between vertexes d.sub.ij by the number of vertexes was explained; however, it was found that the estimation accuracy of 5 mm or less was obtained by defining the feature amount data x.sub.i by a difference between vertexes d.sub.ij of approximately 5% or more. It is preferable to define the feature amount data x.sub.i with the difference between vertexes d.sub.ij of approximately 10% or more, and it is more preferable to define the feature amount data x.sub.i with the difference between vertexes d.sub.ij of approximately 15% or more for obtaining stable estimation accuracy.
Application Example
(194) By integrating the above-described estimation system 1 (refer to
(195) For example, the output screen 300 shown in
Exemplary Embodiment 2
(196) Exemplary embodiment 1 described estimation of deformation of a single organ from the pre-deformation medical image data; however, here, description will be given of a case in which deformation or movement in the positions of multiple organs is to be estimated.
(197)
(198) The example in
(199) The shape and position of organs in the body also change depending on breathing, posture and physical condition in the measurement, and the position moves; however, use of the technique described in Exemplary embodiment 1 makes it possible to estimate the shape and position changes of multiple organs (including tumors) with high accuracy. In other words, in the exemplary embodiment, movement of the position of organs is also learned.
(200) Changes in the shape and position of multiple organs (including tumors) due to breathing, posture and physical condition in the measurement are also examples of deformations by time variation.
(201) In
(202) The capability of estimating the range of deformations and position changes of the multiple organs with high accuracy can be applied to radiation planning in the radiotherapy.
Exemplary Embodiment 3
(203) In Exemplary embodiment 1, the learned model is provided with the feature amount data generated based on the pre-deformation measurement data to estimate the post-deformation shape; however, it may be possible to estimate the post-deformation shape by providing the feature amount data to the learned model that has learned about individual organs or multiple organs, and to check the estimated shape with medical image data, to thereby automatically extract the position of the organs included in the medical image data by using the image recognition technique.
(204)
(205) The deformation estimation device 20 can provide the post-deformation shape estimated by the deformation estimation unit 202 to an image recognition unit 205 to check thereof with the medical image data, and can extract the region of the organ to be the target from the medical image data.
Exemplary Embodiment 4
(206) In the above-described exemplary embodiments, description was given of the case in which both measurement data and learned data were three-dimensional data, but there is a possibility that the measurement data is two-dimensional data. For example, the measurement data is given as endoscopic image data.
(207)
(208) In
(209) H. Fan et al., “A point set generation network for 3D object reconstruction from a single image,” CVPR, 2017.
(210) Note that the estimation processing by the three-dimensional shape estimation unit 60 can be achieved by combining the extraction processing of specific organ regions by graph-cutting with the estimation processing using the learned model that has learned by applying machine learning, deep learning, convolutional neural network, and so on.
(211) The learned model shown here has learned and stored the feature amount data x.sub.i including the difference between each micro region and other micro regions d.sub.ij in the shape model M.sub.(i)I generated from the measurement data of the organ region before deformation, and the relation providing the displacement amount data y.sub.i from each micro region in the shape model M.sub.(i)I of the organ region before deformation to the each corresponding micro region in the shape model M.sub.(i)D of the organ region after deformation.
(212) Use of the technique of the exemplary embodiment makes it possible not only to estimate the post-deformation shape, but also to estimate the position of the tumor that cannot be seen from the appearance, even in the situation where only two-dimensional image data can be measured.
Exemplary Embodiment 5
(213) In the above-described exemplary embodiments, examples of the candidate for the feature amount data x.sub.i included the coordinates v.sub.i of each vertex i, the gradient or the normal vector n.sub.i defining the micro region, the curvature (second derivative, discrete Laplacian) l.sub.i, the difference between vertexes d.sub.ij, the shrinkage ratio s.sub.i of the micro region, and Voronoi area r.sub.i in the micro region; however, as the candidate for the feature amount data x.sub.i, the deviation amount u.sub.i between each micro region and a corresponding region of the average shape model (hereinafter referred to as “average shape model”) may be additionally included. In this case, the deviation amount u.sub.i may be used alone as the feature quantity data x.sub.i, or in combination with other candidates as the feature amount data x.sub.i.
(214)
(215)
(216) Note that it is also possible to provide the average shape model M here as the template shape T and to repeatedly calculate the shape model M.sub.(i) corresponding to the shape data S.sub.(i).
Exemplary Embodiment 6
(217) In the above-described exemplary embodiments, description was given of the case in which the relation between the feature amount data x.sub.i and the displacement amount data y.sub.i before and after the deformation for the same organ was learned, and the post-deformation shape of any measurement data was estimated by using the learned model, which was the result of learning; however, it may be possible to learn the relation between the deviation amount of an object 1 from the average model and the coordinates vs of each vertex s constituting another object 2 having a constraint with the object 1, and to estimate the position of the object 2 from the measurement data of any object 1 by using the learned model as a result of the learning.
(218)
(219) The machine learning device 310 includes a deviation amount calculation unit 310A calculating and outputting a deviation amount u.sub.i per each vertex, and a machine learning unit 310B learning the relation between each deviation amount u.sub.i and the coordinates vs of each vertex s of object 2. In the case of the example, the machine learning unit 310B outputs the corresponding relation function g as a result of the learning.
(220) For example, the object 1 is the liver and the object 2 is the pancreas. The constraint relation here is, for example, adjacent or connected relation. In the case where the object 1 is the liver, the adjacent objects 2 are the stomach, duodenum, and right kidney; the connecting objects 2 are the duodenum, pancreas, spleen, and so on.
(221)
(222) The deformation estimation device 320 includes: a feature amount calculation unit 321 calculating, as the feature amount data, the deviation amount u.sub.i of each vertex i constituting the shape model generated from the measurement data of the object 1 from the average shape model M; an object position estimation unit 322 applying the calculated deviation amount u.sub.i to the learned model to estimate the position of the object 2; and an estimation result output unit 323 outputting the estimation results.
(223) Similar to the above-described exemplary embodiments, the position estimation device 320 is also implemented through execution of programs by a computer.
Exemplary Embodiment 7
(224) In the above exemplary embodiments, description was given of the learning of deformation of an entire object using image data captured or measured the entire object; however, the following description will be given of learning of deformation of an entire object using image data captured or measured a part of the object.
(225) In the following description, three-dimensional image data generated by cone-beam CT is used as an example of the image data captured or measured a part of an object. The cone-beam CT can generate three-dimensional image data of a subject by simply rotating once a movable device configured with an irradiator irradiating the subject with a beam of X-rays in a cone shape or quadrangular-pyramid shape and a two-dimensional detector detecting X-rays having passed through the subject.
(226) For this reason, the cone-beam CT is used, for example, to capture an image of a lung during surgery. However, the volume of lung parenchyma included in the image-capturing scope of the cone-beam CT is less than half the overall lung volume in some cases, and it is not clear which region is captured.
(227)
(228) In the figure, the CT image P1, which has the largest area shown by a mesh, is the CT image captured before the surgery. Unlike the cone-beam CT capturing only a partial image of the lung, the entire lung image is captured. Since the lung is before surgery, the lung is in the aerated condition. The CT image P1 here corresponds to the pre-deformation image I.sub.(i)I in Exemplary embodiment 1.
(229) In the figure, the CT image P2 having the second largest area with a light color is the CT image captured with the cone-beam CT during the surgery. The CT image P2 shows a part of the lung that is in the aerated condition. For any of the specimens A, B, and C, the CT image P2 is a smaller area than the CT image P1.
(230) In the figure, the CT image P3 having the smallest area with a dark color is also the CT image captured with the cone-beam CT during the surgery. The CT image P3 shows a part of the lung that is in the deaerated condition. Consequently, in any of the specimens A, B, and C, the CT image P3 is a smaller region than the CT image P2.
(231) As described above, in the exemplary embodiment, description will be given of the case in which the CT images P2 and P3 are used to learn deformation of the entire lung. However, mere learning of deformation between the CT images P2 and P3, which are the partial images, is performed, the results of the learning cannot be reflected in the learning of the deformation of the entire lung.
(232) Hence, in the exemplary embodiment, deformation between the partial images is learned as variation in a part of the entire lung by alignment with the CT image P1 of the entire lung that was captured before surgery.
(233)
(234) Note that the hardware configuration and the functional configuration of the machine learning device are similar to those of Exemplary embodiment 1. In other words, the machine learning device in Exemplary embodiment 7 also generates a shape model for each of the pre-deformation image I.sub.(i)I and the post-deformation image I.sub.(i)D, and thereafter, upsamples the generated shape model and then executes learning per micro region unit.
(235) In the case of
(236) First, the sampling unit 101C samples each of the pre-deformation image I.sub.(i)IT, the pre-deformation partial image I.sub.(i)IP, and the post-deformation partial image I.sub.(i)DP to generate shape data of the voxel mesh structure or the polygon mesh structure. Here, the shape data corresponding to the pre-deformation image I.sub.(i)IT is referred to as S.sub.(i)IT, the shape data corresponding to the pre-deformation partial image I.sub.(i)IP is referred to as S.sub.(i)IP, and the shape data corresponding to the post-deformation partial image I.sub.(i)D is referred to as S.sub.(i)D.
(237) As described in Exemplary embodiment 1, the size of the pre-deformation image and the post-deformation image also differs depending on differences in the specimens and the image capturing conditions. In addition, the size of the pre-deformation image and the post-deformation image differs between the CT used for pre-surgical image capturing and the cone-beam CT used during surgery. In addition, there are limitations to unify the position and physical condition of the specimens during image capturing.
(238) For this reason, as described in Exemplary embodiment 1, the number of vertexes of the voxels included in the independently-generated shape data is inconsistent in general.
(239) Consequently, in the case of the exemplary embodiment, the generated shape data is aligned with the template shape T to generate the shape model M with the aligned number of vertexes. In
(240) First, by using the template shape T corresponding to the entire right lung, the alignment unit 101D aligns the number of vertexes in the pre-deformation shape data S.sub.(i)IT corresponding to the entire right lung with the same number of vertexes as the template shape T. The details of the processing are the same as those of Exemplary embodiment 1. For example, the pre-deformation shape model M.sub.(i)IT with 500 vertexes is generated.
(241) This alignment is the alignment between the global coordinate systems.
(242) Next, the alignment unit 101D uses the generated pre-deformation shape model M.sub.(i)IT corresponding to the entire right lung as a template shape to align the pre-deformation partial shape data S.sub.(i)IP corresponding to a part of the right lung of the same specimen. As a result, the pre-deformation partial shape model M.sub.(i)IP corresponding to the pre-deformation partial shape data S.sub.(i)IP is generated. Of course, the number of vertexes constituting the pre-deformation partial shape model M.sub.(i)IP is less than the number of vertexes in the template shape.
(243) The site and volume to be captured by the cone-beam CT are, as shown in
(244) This alignment is the alignment between the global coordinate system and the local coordinate system.
(245) Next, the alignment unit 101D uses the generated pre-deformation shape model M.sub.(i)IP corresponding to a part of the right lung in the aerated condition as a template shape to align the post-deformation partial shape data S.sub.(i)DP corresponding to a part of the right lung in the deaerated condition of the same specimen. As a result, the post-deformation partial shape model M.sub.(i)DP corresponding to the post-deformation partial shape data S.sub.(i)DP is generated. However, the number of vertexes constituting the post-deformation partial shape model M.sub.(i)DP is the same as the number of vertexes of the pre-deformation partial shape model M.sub.(i)IP used as the template shape.
(246) This alignment is the alignment between the local coordinate systems.
(247)
(248) In the case of the specimen A shown in
(249) The degree of inclusion (dice similarity distance: DSC) between the estimated values and true values of the vertex group is 98.93%.
(250) The MD of the specimen B is 0.15 mm, the HD is 0.79 mm, and the DSC is 99.24%. The MD of the specimen C is 0.23 mm, the HD is 1.30 mm, and the DSC is 98.94%.
(251) As shown in
(252)
(253) In the case of
(254) As shown in
(255)
(256) The pre-deformation partial shape model group M.sub.(i1)IP, . . . , M.sub.(ik)IP is a model group generated by upsampling the pre-deformation partial shape model M.sub.(i)IP. The post-deformation partial shape model group M.sub.(i1)DP, . . . , M.sub.(ik)DP is a model group generated by upsampling the post-deformation partial shape model M.sub.(i)DP.
(257) In the case of Exemplary embodiment 1, the feature amount per vertex unit is calculated directly from the model group generated by upsampling; however, in the exemplary embodiment, the vertexes near the boundary are removed in advance by using near-boundary removal units 103A0 and 103B0. The partial images captured with cone-beam CT have low reliability near the boundary thereof with the region of the entire lung that was not captured.
(258)
(259) In the case of
(260) On the other hand, in the specimen B, the upper half of the lung is captured with the cone-beam CT, but the lower half of the lung is not captured. Consequently, the boundary appears only at one location: The lower end side of the CT image P2 (pre-deformation partial shape model) and the CT image P3 (post-deformation partial shape model).
(261) In the exemplary embodiment, the range of 10 mm from each boundary is excluded from the subject to be learned. Note that the value is merely an example. However, the vertexes to be excluded from the subject to be learned may be determined by the ratio (%) of the number of vertexes constituting the shape model, not by the distance from the boundary. For example, 10% can be used.
(262) Let us return to the description of
(263) On the other hand, the near-boundary removal unit 103B0 removes the vertexes existing near the boundary from each of the models in the post-deformation partial shape model group M.sub.(i1)DP, . . . , M.sub.(ik)DP. The group of remaining vertexes (namely, the group of vertexes to be learned) is outputted to the displacement calculation unit 103B.
(264)
(265)
(266) In all the figures, the length of the displacement amount data y.sub.i is shorter than it is in the actuality, taking the visibility into consideration.
(267) The displacement amount data y.sub.i representing the displacement before and after the deformation of each vertex to be learned is calculated by the displacement calculation unit 103B (refer to
(268) The calculated feature amount data x.sub.i and displacement amount data y.sub.i are given to the machine learning unit 103C (refer to
(269)
(270) As shown in
(271)
(272) The vertical axis in the figure indicates the changes in lung volume before and after surgery, and the horizontal axis indicates the specimen.
(273) Note that, in the above description, the cone-beam CT is used for acquiring the partial image; however, the device for acquiring the partial image is not limited to the cone-beam CT.
Other Exemplary Embodiments
(274) So far, the exemplary embodiments according to the present invention have been described, but the technical scope of the present invention is not limited to the scope described in the above-described exemplary embodiments.
(275) It is obvious from the following claims that various modifications and improvements added to the above-described exemplary embodiment are also included in the technical scope of the present invention.
(276) For example, in the above exemplary embodiments, a single vertex is assumed as the micro region, but multiple vertexes may be included.
(277) Moreover, in the above exemplary embodiments, the learned model is generated by learning the relation between the feature amount data and the displacement amount data for each vertex by using the kernel regression model; however, it may also be possible to generate the learned model by using the deep learning or the convolutional neural network.
(278) In addition, in the above exemplary embodiments, the relation of distance between the vertexes constituting the shape model is not particularly defined, but may be constrained to be equally spaced.
(279) In the above exemplary embodiments, the learned model is generated by learning the organ before and after deformation exclusively based on the medical image data; however, the technology according to the present invention is not limited to the medical image data.
(280) In the example shown in
REFERENCE SIGNS LIST
(281) 1: Estimation system 10: Machine learning device 20: Deformation estimation device 30: Database 40: Network 50: Terminal 60: Three-dimensional shape estimation unit 101: Shape model generation unit 101A: Sampling unit 101B: Alignment unit 102: Model upsampling unit 103: Deformation learning unit 103A: Feature amount calculation unit 103B: Displacement calculation unit 103C: Machine learning unit 201: Feature amount calculation unit 202: Deformation estimation unit 203: Estimation result output unit 205: Image recognition unit 300: Output screen 301: Observation image section 302: Estimation image section