DYNAMIC TRACKING METHODS FOR IN-VIVO THREE-DIMENSIONAL KEY POINT AND IN-VIVO THREE-DIMENSIONAL CURVE

20240005532 ยท 2024-01-04

    Inventors

    Cpc classification

    International classification

    Abstract

    Dynamic tracking methods for an in-vivo three-dimensional key point and an in-vivo three-dimensional curve can include mapping a first local region to a first local point cloud and mapping a second local region to a second local point cloud according to a mapping relation between an endoscopic image and the point clouds; determining a first three-dimensional key point of a first two-dimensional key point on the first local point cloud, and acquiring a second three-dimensional key point on the second local point cloud through a coordinate transformation; and mapping the second three-dimensional key point back to the second local region, so as to acquire a second two-dimensional key point from a next image; and acquiring two-dimensional coordinates of a tracked key point by minimizing a preset optimization function in combination with an initial two-dimensional key point, and finally acquiring corresponding three-dimensional coordinates.

    Claims

    1-13. (canceled)

    14. A minimally invasive key site navigation oriented dynamic tracking method for an in-vivo three-dimensional key point, comprising: S11, reading an endoscopic image, and acquiring a first two-dimensional key point from a current image according to selection of a doctor; S12, tracking a first local region encompassing the first two-dimensional key point on the current image, acquiring a second local region from a next image, and determining an initial two-dimensional key point of the first two-dimensional key point on the next image; S13, mapping the first local region to a first local point cloud and mapping the second local region to a second local point cloud according to a mapping relation between the endoscopic image and the point clouds, determining a first three-dimensional key point of the first two-dimensional key point on the first local point cloud, and acquiring a second three-dimensional key point on the second local point cloud through a coordinate transformation; and S14, mapping the second three-dimensional key point back to the second local region, so as to acquire a second two-dimensional key point from the next image, acquiring two-dimensional coordinates of a tracked key point by minimizing a preset optimization function in combination with the initial two-dimensional key point, and finally acquiring corresponding three-dimensional coordinates.

    15. The tracking method for an in-vivo three-dimensional key point according to claim 14, wherein S12 comprises: S121, defining a k1th image as I(k1)custom-character.sup.WH3, wherein W denotes a width of the endoscopic image, H denotes a height of the endoscopic image, I(k) denotes a kth image, and p(k1) denotes a first two-dimensional key point, coordinates of which are (u.sub.1, v.sub.1); and taking the first two-dimensional key point p(k1) as a center p.sup.c(k1) of the first local region, and determining a first local region R(k1) according to a preset region shape and side length; S122, performing feature matching on feature points of images I(k31 1), I(k) through an optical flow method, and acquiring a center p.sup.c(k), corresponding to p.sup.c(k1), of the second local region from the image I(k), p c ( k ) = p c ( k - 1 ) + 1 m .Math. i = 1 m ( p i f ( k ) - p i f ( k - 1 ) ) wherein p.sup.f(k)={p.sub.1.sup.f(k), p.sub.2.sup.f(k), . . . , p.sub.m.sup.f(k)} denotes a feature point on the image I(k), and m denotes the number of the feature point on the image I(k); S123, determining a second local region custom-character(k) according to the center p.sup.c(k) and the preset region shape and side length; and S124, directly determining an initial two-dimensional key point p.sup.o(k) of the first two-dimensional key point p(k1) on the next image still through the optical flow method.

    16. The dynamic tracking method for an in-vivo three-dimensional key point according to claim 15, wherein S13 comprises: S131, estimating a depth of the endoscopic image to obtain a depth image corresponding to the endoscopic image, acquiring space information and color information of each pixel from the depth image and the endoscopic image respectively through reading in rows, and acquiring a first local point cloud custom-character(k1) and a second local point cloud custom-character(k); S132, determining a first three-dimensional key point P(k1)custom-character.sup.3 of the first two-dimensional key point p(k1) on the first local point cloud custom-character(k1),
    P(k1)=(p(k1)) wherein denotes a mapping relation from custom-character(k1) to custom-character(k 1), which is recorded as custom-character(k1).fwdarw.custom-character(k 1); S133, acquiring a feature point pair of the local regions custom-character(k1), custom-character(k) through the optical flow method and recording same as X and Y, respectively, so that X and Y are in a coordinate transformation relation: Y = [ A t ] [ X 1 ] wherein Acustom-character.sup.33, tcustom-character.sup.31, =[A t].sup.Tcustom-character.sup.43 denote parameters of a fitting function, is acquirable from the following formula through least squares:
    =([X1].sup.T[X1]).sup.1[X1].sup.TY and a transformation matrix of an affine transformation of P(k1), P(k) is: T A = [ A t 0 T 1 ] 4 4 wherein 0.sup.T=(0,0,0); and S134, performing a three-dimensional affine transformation on the first three-dimensional key point P(k1), wherein a matrix form is:
    P(k)=T.sub.AP(k1) and searching for a nearest point from the second local point cloud custom-character(k) to obtain an initial position of a second three-dimensional key point P(k).

    17. The dynamic tracking method for an in-vivo three-dimensional key point according to claim 14, wherein the optimization function in S14 is as follows:
    custom-character=2custom-character.sub.siftcustom-character.sub.optical wherein custom-character denotes the optimization function; custom-character.sub.sift denotes a cosine similarity of a scale invariant feature transform (SIFT) feature vector: J sift ( u , v , k ) = ( p ( k - 1 ) ) T ( p u + u , v + v ( k ) ) .Math. ( p ( k - 1 ) ) .Math. .Math. ( p u + u , v + v ( k ) ) .Math. wherein (p(k1)) denotes a feature descriptor of the first two-dimensional key point p(k1) and is a vector, .Math. denotes a norm of the vector, (u, v) denote coordinates of the second two-dimensional key point p(k), p.sub.u+u,v+v(k) denotes a neighborhood point of the second two-dimensional key point p(k), u and v are coordinate offset of the second two-dimensional key point p(k), (p.sub.u+u,v+v(k)) denotes a feature descriptor of the neighborhood point p.sub.u+u,v+v(k) and is a vector, and custom-character.sub.optical denotes an effect of optical flow information: J optical ( u , v , k ) = a T b .Math. a .Math. .Math. b .Math. wherein
    a=(p.sub.u+u,v+v(k)p(k)).sup.T,b=(p.sup.o(k)p(k)).sup.T.

    18. The dynamic tracking method for an in-vivo three-dimensional key point according to claim 15, wherein the optimization function in S14 is as follows:
    custom-character=2custom-character.sub.siftcustom-character.sub.optical wherein custom-character denotes the optimization function; custom-character.sub.sift denotes a cosine similarity of a scale invariant feature transform (SIFT) feature vector: J sift ( u , v , k ) = ( p ( k - 1 ) ) T ( p u + u , v + v ( k ) ) .Math. ( p ( k - 1 ) ) .Math. .Math. ( p u + u , v + v ( k ) ) .Math. wherein (p(k1)) denotes a feature descriptor of the first two-dimensional key point p(k1) and is a vector, .Math. denotes a norm of the vector, (u, v) denote coordinates of the second two-dimensional key point p(k), p.sub.u+u,v+v(k) denotes a neighborhood point of the second two-dimensional key point p(k), u and v are coordinate offset of the second two-dimensional key point p(k), (p.sub.u+u,v+v(k)) denotes a feature descriptor of the neighborhood point p.sub.u+u,v+v(k) and is a vector, and custom-character.sub.optical denotes an effect of optical flow information: J optical ( u , v , k ) = a T b .Math. a .Math. .Math. b .Math. wherein
    a=(p.sub.u+u,v+v(k)p(k)).sup.T,b=(p.sup.o(k)p(k)).sup.T.

    19. The dynamic tracking method for an in-vivo three-dimensional key point according to claim 16, wherein the optimization function in S14 is as follows:
    custom-character=2custom-character.sub.siftcustom-character.sub.optical wherein custom-character denotes the optimization function; custom-character.sub.sift denotes a cosine similarity of a scale invariant feature transform (SIFT) feature vector: J sift ( u , v , k ) = ( p ( k - 1 ) ) T ( p u + u , v + v ( k ) ) .Math. ( p ( k - 1 ) ) .Math. .Math. ( p u + u , v + v ( k ) ) .Math. wherein (p(k1)) denotes a feature descriptor of the first two-dimensional key point p(k1) and is a vector, .Math. denotes a norm of the vector, (u, v) denote coordinates of the second two-dimensional key point p(k), p.sub.u+u,v+v(k) denotes a neighborhood point of the second two-dimensional key point p(k), u and v are coordinate offset of the second two-dimensional key point p(k), (p.sub.u+u,v+v(k)) denotes a feature descriptor of the neighborhood point p.sub.u+u,v+v(k) and is a vector, and custom-character.sub.optical denotes an effect of optical flow information: J optical ( u , v , k ) = a T b .Math. a .Math. .Math. b .Math. wherein
    a=(p.sub.u+u,v+v(k)p(k)).sup.T,b=(p.sup.o(k)p(k)).sup.T.

    20. The dynamic tracking method for an in-vivo three-dimensional key point according to claim 17, wherein the acquiring two-dimensional coordinates of a tracked key point by minimizing a preset optimization function, and finally acquiring corresponding three-dimensional coordinates in S14 comprises: acquiring and {circumflex over (v)} by traversing and searching for u and v, so as to satisfy the following expression: ( u ( k ) , v ( k ) ) = argmin ( u , v ) ( J ( u , v , k ) ) and acquiring two-dimensional coordinates of a tracked key point p.sub.u+u,v+v(k) after ideal offset ((k), {circumflex over (v)}(k)) is obtained, and then finally acquiring corresponding three-dimensional coordinates according to the mapping relation between the endoscopic image and the point clouds.

    21. An electronic apparatus, comprising: one or more processors; a memory; and one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the one or more processors, and the program comprises executing the dynamic tracking method for an in-vivo three-dimensional key point according to claim 14.

    22. The electronic apparatus according to claim 21, wherein S12 comprises: S121, defining a k1th image as I(k1)custom-character.sup.WH3 wherein W denotes a width of the endoscopic image, H denotes a height of the endoscopic image, I(k) denotes a kth image, and p(k1) denotes a first two-dimensional key point, coordinates of which are (u.sub.1, v.sub.1); and taking the first two-dimensional key point p(k1) as a center p.sup.c(k1) of the first local region, and determining a first local region custom-character(k1) according to a preset region shape and side length; S122, performing feature matching on feature points of images I(k1), I(k) through an optical flow method, and acquiring a center p.sup.c(k), corresponding to p.sup.c(k1), of the second local region from the image I(k), p c ( k ) = p c ( k - 1 ) + 1 m .Math. i = 1 m ( p i f ( k ) - p i f ( k - 1 ) ) wherein p.sup.f(k)={p.sub.1.sup.f(k), p.sub.2.sup.f(k), . . . , p.sub.m.sup.f(k)} denotes a feature point on the image I(k), and m denotes the number of the feature point on the image I(k); S123, determining a second local region custom-character(k) according to the center p.sup.c(k) and the preset region shape and side length; and S124, directly determining an initial two-dimensional key point p.sup.o(k) of the first two-dimensional key point p(k1) on the next image still through the optical flow method.

    23. The electronic apparatus according to claim 22, wherein S13 comprises: S131, estimating a depth of the endoscopic image to obtain a depth image corresponding to the endoscopic image, acquiring space information and color information of each pixel from the depth image and the endoscopic image respectively through reading in rows, and acquiring a first local point cloud custom-character(k1) and a second local point cloud custom-character(k); S132, determining a first three-dimensional key point P(k1)custom-character.sup.3 of the first two-dimensional key point p(k1) on the first local point cloud custom-character(k1),
    P(k1)=(p(k1)) wherein denotes a mapping relation from custom-character(k1) to custom-character(k1), which is recorded as custom-character(k1).fwdarw.custom-character(k1); S133, acquiring a feature point pair of the local regions custom-character(k1), custom-character(k) through the optical flow method and recording same as X and Y, respectively, so that X and Y are in a coordinate transformation relation: Y = [ A t ] [ x 1 ] wherein Acustom-character.sup.33, tcustom-character.sup.31, =[A t].sup.Tcustom-character.sup.43 denote parameters of a fitting function, is acquirable from the following formula through least squares:
    =[X1].sup.T[X1]).sup.1[X1].sup.TY and a transformation matrix of an affine transformation of P(k1), P(k) is: T A = [ A t 0 T 1 ] 4 4 wherein 0.sup.T=(0,0,0); and S134, performing a three-dimensional affine transformation on the first three-dimensional key point P(k1), wherein a matrix form is:
    P(k)=T.sub.AP(k1) and searching for a nearest point from the second local point cloud custom-character(k) to obtain an initial position of a second three-dimensional key point P(k).

    24. The electronic apparatus according to claim 21, wherein the optimization function in S14 is as follows:
    custom-character=2custom-character.sub.siftcustom-character.sub.optical wherein custom-character denotes the optimization function; custom-character.sub.sift denotes a cosine similarity of a scale invariant feature transform (SIFT) feature vector: J sift ( u , v , k ) = ( p ( k - 1 ) ) T ( p u + u , v + v ( k ) ) .Math. ( p ( k - 1 ) ) .Math. | .Math. ( p u + u , v + v ( k ) ) .Math. wherein (p(k1)) denotes a feature descriptor of the first two-dimensional key point p(k1) and is a vector, .Math. denotes a norm of the vector, (u, v) denote coordinates of the second two-dimensional key point p(k), p.sub.u+u,v+v(k) denotes a neighborhood point of the second two-dimensional key point p(k), u and v are coordinate offset of the second two-dimensional key point p(k), (p.sub.u+u,v+v(k)) denotes a feature descriptor of the neighborhood point p.sub.u+u,v+v(k) and is a vector, and custom-character.sub.optical denotes an effect of optical flow information: J optical ( u , v , k ) = a T b .Math. a | .Math. .Math. b .Math. wherein
    a=(p.sub.u+u,v+v(k)p(k)).sup.T,b=(p.sup.o(k)p(k)).sup.T.

    25. The electronic apparatus according to claim 24, wherein the acquiring two-dimensional coordinates of a tracked key point by minimizing a preset optimization function, and finally acquiring corresponding three-dimensional coordinates in S14 comprises: acquiring and {circumflex over (v)} by traversing and searching for u and v, so as to satisfy the following expression: ( v ( k ) ) = arg min ( u , v ) ( J ( u , v , k ) ) and acquiring two-dimensional coordinates of a tracked key point p.sub.u+u,v+v(k) after ideal offset ((k), {circumflex over (v)}(k)) is obtained, and then finally acquiring corresponding three-dimensional coordinates according to the mapping relation between the endoscopic image and the point clouds.

    26. A minimally invasive key trajectory navigation oriented dynamic tracking method for an in-vivo three-dimensional curve, comprising: S21, reading an endoscopic image, acquiring an operation path curve from a current image according to selection of a doctor, and acquiring a plurality of first two-dimensional key points through which the operation path curve passes; S22, tracking a first local region encompassing the first two-dimensional key point on the current image, and acquiring a second local region from a next image; S23, mapping the first local region to a first local point cloud and mapping the second local region to a second local point cloud according to a mapping relation between the endoscopic image and the point clouds, determining a first three-dimensional key point of the first two-dimensional key point on the first local point cloud, and acquiring a second three-dimensional key point on the second local point cloud through a coordinate transformation; S24, reducing a dimension of the first local point cloud to obtain a first two-dimensional point cloud, and acquiring a second two-dimensional key point of the first three-dimensional key point on the first two-dimensional point cloud; reducing a dimension of the second local point cloud to obtain a second two-dimensional point cloud, and acquiring a third two-dimensional key point of the second three-dimensional key point on the second two-dimensional point cloud; and acquiring two-dimensional coordinates of a tracked key point on the two-dimensional point cloud by minimizing a preset optimization function according to the second two-dimensional key point and the third two-dimensional key point; and S25, acquiring three-dimensional coordinates of each tracked key point according to a mapping relation between the point clouds before and after dimension reduction, performing curve fitting, and finally obtaining a three-dimensional curve by means of tracking.

    27. The dynamic tracking method for an in-vivo three-dimensional curve according to claim 26, wherein S22 comprises: S221, defining a k1th image a I/(k1)custom-character.sup.WH3, wherein W denotes a width of the endoscopic image, H denotes a height of the endoscopic image, I(k) denotes a kth image, and custom-character(k1)={p.sub.1(k1), . . . , p.sub.n(k1)} denote a plurality of first two-dimensional key points; and determining maximums and minimums of all the first two-dimensional key points on a u axis and a v axis of an image coordinate system, respectively, selecting a position of a median on each axis as p.sup.c(k1)=(u.sub.1, v.sub.1)R.sup.2, and determining a first local region custom-character(k1) according to a preset region shape and side length; S222, performing feature matching on feature points of images I(k1), I(k) through an optical flow method, and acquiring a center p.sup.c(k), corresponding to p.sup.c(k1), of the second local region from the image I(k), p c ( k ) = p c ( k - 1 ) + 1 m .Math. i = 1 m ( p i f ( k ) - p i f ( k - 1 ) ) where p.sup.f(k)={p.sub.1.sup.f(k), p.sub.2.sup.f(k), . . . , p.sub.m.sup.f(k)} denotes a feature point on the image I(k), and m denotes the number of the feature point on the image I(k); and S223, determining a second local region custom-character(k) according to the center p.sup.c(k) and the preset region shape and side length.

    28. The dynamic tracking method for an in-vivo three-dimensional curve according to claim 27, wherein S23 comprises: S231, estimating a depth of the endoscopic image to obtain a depth image corresponding to the endoscopic image, acquiring space information and color information of each pixel from the depth image and the endoscopic image respectively through reading in rows, and acquiring a first local point cloud custom-character(k1) and a second local point cloud custom-character(k); S232, determining a first three-dimensional key point custom-character(k1)={P.sub.1(k1), . . . , P.sub.n(k1)}custom-character(k1), P.sub.i(k1)custom-character.sup.3 of the first two-dimensional key point custom-character(k1) on the first local point cloud custom-character(k1),
    P.sub.i(k1)=(p.sub.i(k1)) wherein denotes a mapping relation from custom-character(k1) to custom-character(k1), which is recorded as custom-character(k1).fwdarw.custom-character(k1); S233, acquiring a feature point pair of the local regions custom-character(k1), custom-character(k) through the optical flow method and recording same as X and Y, respectively, so that X and Y are in a coordinate transformation relation: Y = [ A t ] [ X 1 ] [ x1 ] wherein Acustom-character.sup.33, tcustom-character.sup.31, =[A t].sup.Tcustom-character.sup.43 denote parameters of a fitting function, is acquirable from the following formula through least squares:
    =([X1].sup.T[X1]).sup.1[X1].sup.TY and a transformation matrix of an affine transformation of P(k1), P(k) is: T A = [ A t 0 T 1 ] 4 4 wherein 0.sup.T=(0,0,0); and S234, performing a three-dimensional affine transformation on the first three-dimensional key point P(k1), and searching for a nearest point from the second local point cloud custom-character(k) to obtain an initial position of a second three-dimensional key point P(k), wherein
    P(k)=T.sub.AP(k1).

    29. The dynamic tracking method for an in-vivo three-dimensional curve according to claim 26, wherein the optimization function in S24 is as follows:
    custom-character=1custom-character.sub.sift+custom-character.sub.shape where custom-character denotes the optimization function; custom-character.sub.sift denotes a cosine similarity of an SIFT feature vector: J sift ( T i ( k ) ) = ( T i ( k - 1 ) ) T ( T i ( k ) ) .Math. ( T i ( k - 1 ) ) .Math. .Math. ( T i ( k ) ) .Math. wherein (T.sub.i(k1)) denotes a feature descriptor of an ith second two-dimensional key point T.sub.i(k1) and is a vector, (T.sub.i(k)) denotes a feature descriptor of a neighborhood point of an ith third two-dimensional key point T.sub.i(k) and is a vector, and .Math. denotes a norm of the vector; and custom-character.sub.shape denotes a difference in cosine values of included angles between adjacent key points on different curves: s h a p e ( T i ( k ) ) = 1 n .Math. i = 0 n .Math. g ( T i ( k - 1 ) ) - g ( T i ( k ) ) .Math. "\[RightBracketingBar]" wherein g(T.sub.i(k)) is the cosine value of the included angle, which is specifically calculated as follows: g ( T i ( k ) ) = a T b .Math. a .Math. .Math. b .Math.
    wherein a=(T.sub.i+1(k)T.sub.i(k)).sup.T,b=(T.sub.i1(k)T.sub.i(k)).sup.T.

    30. The dynamic tracking method for an in-vivo three-dimensional curve according to any one of claim 27, wherein the optimization function in S24 is as follows:
    custom-character=1custom-character.sub.sift+custom-character.sub.shape wherein custom-character denotes the optimization function; custom-character.sub.sift denotes a cosine similarity of an SIFT feature vector: J sift ( T i ( k ) ) = ( T i ( k - 1 ) ) T ( T i ( k ) ) .Math. ( T i ( k - 1 ) ) .Math. .Math. ( T i ( k ) ) .Math. wherein (T.sub.i(k1)) denotes a feature descriptor of an ith second two-dimensional key point T.sub.i(k1) and is a vector, (T.sub.i(k)) denotes a feature descriptor of a neighborhood point of an ith third two-dimensional key point T.sub.i(k) and is a vector, and .Math. denotes a norm of the vector; and custom-character.sub.shape denotes a difference in cosine values of included angles between adjacent key points on different curves: s h a p e ( T i ( k ) ) = 1 n .Math. i = 0 n | g ( T i ( k - 1 ) ) - g ( T i ( k ) ) | wherein g(T.sub.i(k)) is the cosine value of the included angle, which is specifically calculated as follows: g ( T i ( k ) ) = a T b .Math. a .Math. .Math. b .Math.
    wherein a=(T.sub.i+1(k)T.sub.i(k)).sup.T,b=(T.sub.i1(k)T.sub.i(k)).sup.T.

    31. The dynamic tracking method for an in-vivo three-dimensional curve according to claim 29, wherein the acquiring two-dimensional coordinates of a tracked key point on the two-dimensional point cloud by minimizing a preset optimization function according to the second two-dimensional key point and the third two-dimensional key point in S24 comprises: minimizing custom-character by traversing and searching for neighborhood points of all third two-dimensional key points T.sub.i(k), so as to satisfy:
    {circumflex over (T)}.sub.i=argmin(1custom-character.sub.sift(T.sub.i(k))+custom-character.sub.shape(T.sub.i(k))) and acquiring an ideal key point T.sub.i set by minimizing the optimization function.

    32. The dynamic tracking method for an in-vivo three-dimensional curve according to claim 29, wherein the performing curve fitting, and finally obtaining a three-dimensional curve by means of tracking in S25 comprises: performing interpolation fitting on a line through an equation of a B-spline curve, wherein a general equation of the B-spline curve is P(t)=.sub.i=0.sup.n P.sub.i F.sub.i,k(t): wherein P.sub.i denotes a feature point of a control curve, F.sub.i,k(t) denotes a kth order B-spline basis function, and the three-dimensional curve is tracked through curve interpolation fitting.

    33. The dynamic tracking method for an in-vivo three-dimensional curve according to claim 29, wherein a first two-dimensional key point acquisition process in S21 comprises: defining custom-character.sup.p={p.sub.0.sup.p, . . . , p.sub.j.sup.p, . . . , p.sub.l1.sup.p} as a pixel through which the operation path curve passes; wherein for a point on the curve, a curvature of a jth pixel on the curve is: K j = v j + - v j u j + - u j , j = 0 , , 2 .Math. , j < l - 1 , wherein p.sub.j+.sup.p=[u.sub.j+ v.sub.j+] denotes coordinates of a j+th pixel on the curve, and a denotes an interval number of the pixels when the curvature of the pixel is solved; and for curvatures of two consecutive pixels, when |K.sub.j+K.sub.j|, denotes a curvature threshold; and determining all first two-dimensional key points custom-character={p.sub.1, . . . , p.sub.i, . . . , p.sub.n} in combination with p.sub.j+.sup.p as a key point on the operation path curve and a start point and an end point of the curve, wherein n denotes a total number of the first two-dimensional key points.

    Description

    BRIEF DESCRIPTION OF THE DRAWINGS

    [0028] In order to describe the technical solutions in the embodiments of the present disclosure or in the prior art more clearly, the accompanying drawings required for describing the embodiments or the prior art are briefly described below. Apparently, the accompanying drawings in the following description show merely some embodiments of the present disclosure. Those of ordinary skill in the art can still derive other accompanying drawings from these accompanying drawings without creative efforts.

    [0029] FIG. 1 is a schematic flowchart of a minimally invasive key site navigation oriented dynamic tracking method for an in-vivo three-dimensional key point according to Embodiment 1 of the present disclosure;

    [0030] FIG. 2 is a relation diagram among an initial two-dimensional key point, a second two-dimensional key point, and a neighborhood point of the second two-dimensional key point according to Embodiment 1 of the present disclosure; and

    [0031] FIG. 3 is a schematic flowchart of a minimally invasive key trajectory navigation oriented dynamic tracking method for an in-vivo three-dimensional curve according to Embodiment 2 of the present disclosure.

    DETAILED DESCRIPTION OF THE EMBODIMENTS

    [0032] In order to make the objectives, technical solutions, and advantages in the embodiments of the present disclosure clearer, the technical solutions in the embodiments of the present disclosure are described clearly and completely. Apparently, the described embodiments are some embodiments rather than all embodiments of the present disclosure. All other embodiments derived by those of ordinary skill in the art based on the embodiments of the present disclosure without creative efforts fall within the scope of protection of the present disclosure.

    [0033] Embodiments of the present disclosure provide dynamic tracking methods for an in-vivo three-dimensional key point and an in-vivo three-dimensional curve. Accordingly, the technical problems that a three-dimensional key point cannot be precisely located, and an operation path curve cannot be tracked in real time are solved.

    [0034] The technical solutions in the embodiments of the present disclosure are intended to solve the above technical problems. A general idea is as follows:

    [0035] A minimally invasive key site navigation oriented dynamic tracking method for an in-vivo three-dimensional key point according to an embodiment of the present disclosure is configured to dynamically track a manual site on the basis of a three-dimensional point cloud in a robot based tele-operation and mainly applied to, but not limited to, minimally invasive endoscopic operation scenes. The technical solution can be specifically summarized as follows: a doctor selects a key point from an intraoperative image and maps the key point to the three-dimensional point cloud. The key point is determined preliminary through a three-dimensional affine transformation between two point clouds. Then, an optimization function is constructed in combination with a feature descriptor and texture information such as an optical flow. A position of the key point is precisely determined in a neighborhood. Therefore, an endoscopic image oriented three-dimensional key point tracking is realized.

    [0036] Aiming at a flexible and dynamic in-vivo environment, in the embodiment of the present disclosure, the key point is selected manually through the intraoperative image and updated in real time on the three-dimensional point cloud. Therefore, an operation path is updated accurately under a complex and changeable environment. Aiming at an indistinct in-vivo environment feature, the selected key point is tracked initially through the three-dimensional affine transformation, and the three-dimensional key point in the in-vivo environment is precisely and dynamically located and tracked in combination with the texture information and optical flow information.

    [0037] A minimally invasive key trajectory navigation oriented dynamic tracking method for an in-vivo three-dimensional curve according to an embodiment the present disclosure is mainly applied to, but not limited to, minimally invasive endoscopic operation scenes. Accordingly, a tele-operation can be guided more precisely, and an operation robot can complete autonomous manipulation more remarkably.

    [0038] In a scene application, the doctor plans a curved operation path on an intraoperative image upon his/her own knowledge and experience and determines a key point on a curve. A transformation matrix is acquired through a three-dimensional affine transformation between two point clouds. Coordinates of a key point on a source point cloud are transformed through the transformation matrix to obtain an initial position of a three-dimensional key point on a target point cloud. An optimization function is constructed in combination with the texture information of the endoscopic image and shape information of the curve. The three-dimensional key point is precisely located near an initial key point by minimizing the optimization function, and curve fitting is performed to realize dynamic curve fitting on the three-dimensional point cloud.

    [0039] Aiming at a flexible and dynamic in-vivo environment, the operation path is planned manually through the intraoperative images and updated in real time on the three-dimensional point cloud. Therefore, the operation path is updated accurately under a complex and changeable environment. Aiming at an indistinct in-vivo environment feature, the key point on the curve is initially tracked through three-dimensional point cloud registration, and the curve in the in-vivo environment is precisely and dynamically located and tracked in combination with the texture information and the shape information.

    [0040] For a better understanding of the above technical solutions, the above technical solutions are described in detail below with reference to the accompanying drawings and particular embodiments of the description.

    Embodiment 1

    [0041] As shown in FIG. 1, an embodiment of the present disclosure provides a minimally invasive key site navigation oriented dynamic tracking method for an in-vivo three-dimensional key point. The method includes: [0042] S11, an endoscopic image is read, and a first two-dimensional key point is acquired from a current image according to selection of a doctor; [0043] S12, a first local region encompassing the first two-dimensional key point on the current image is tracked, a second local region is acquired from a next image, and an initial two-dimensional key point of the first two-dimensional key point on the next image is determined; [0044] S13, the first local region is mapped to a first local point cloud and the second local region is mapped to a second local point cloud according to a mapping relation between the endoscopic image and the point clouds, a first three-dimensional key point of the first two-dimensional key point on the first local point cloud is determined, and a second three-dimensional key point on the second local point cloud is acquired through a coordinate transformation; and [0045] S14, the second three-dimensional key point is mapped back to the second local region, so as to acquire a second two-dimensional key point from the next image, two-dimensional coordinates of a tracked key point are acquired by minimizing a preset optimization function in combination with the initial two-dimensional key point, and corresponding three-dimensional coordinates are finally acquired.

    [0046] In the embodiment of the present disclosure, a selected key point is initially tracked through a three-dimensional affine transformation. The three-dimensional key point in the in-vivo environment is precisely and dynamically located and tracked in combination with texture information and optical flow information.

    [0047] Each step of the above technical solution will be described in detail below with reference to specific contents:

    [0048] In step S11, the endoscopic image is read, and the first two-dimensional key point is acquired from the current image according to selection of the doctor.

    [0049] In the present step, the doctor labels the two-dimensional key point on an intraoperative image for subsequent display and update of the key point on a three-dimensional point cloud. Accordingly, the information is transmitted intuitively and accurately, and an operation efficiency is improved.

    [0050] Step S12 that a first local region encompassing the first two-dimensional key point on the current image is tracked, a second local region is acquired from a next image, and an initial two-dimensional key point of the first two-dimensional key point on the next image is determined specifically includes: [0051] S121, firstly, a k1th image is defined as I(k1)custom-character.sup.WH3, where W denotes a width of the endoscopic image, H denotes a height of the endoscopic image, I(k) denotes a kth image, and p(k1) denotes a first two-dimensional key point, coordinates of which are (u.sub.1, v.sub.1); and [0052] the first two-dimensional key point p(k1) is taken as a center p.sup.c(k1) of the first local region, and a first local region custom-character(k1) is determined according to a preset region shape and side length; [0053] S122, feature matching is performed on feature points of images I(k1), I(k) through an optical flow method, a movement direction and a distance between two frames are acquired through average differences of pixel coordinates of a feature point pair, and a center, corresponding to p.sup.c(k1), of the second local region on the image I(k) is expressed as:

    [00001] p c ( k ) = p c ( k - 1 ) + 1 m .Math. i = 1 m ( p i f ( k ) - p i f ( k - 1 ) ) [0054] where p.sup.c(k) denotes the center of the second local region, coordinates of which are (u.sub.2, v.sub.2); and


    p.sup.f(k)={p.sub.1.sup.f(k),p.sub.2.sup.f(k), . . . ,p.sub.m.sup.f(k)}

    denotes a feature point on the image I(k), and m denotes the number of the feature point on the image I(k); [0055] S123, a second local region custom-character(k) is determined according to the center p.sup.c(k) and the preset region shape and side length; and [0056] S124, an initial two-dimensional key point p.sup.o(k) of the first two-dimensional key point p(k1) on the next image is directly determined still through the optical flow method.

    [0057] Apparently, the above preset region shape and side length may be selected as actually required and will not be strictly limited herein. Taking a rectangular region custom-character(k1) with a size of LL as an example, p.sup.c(k1) denotes a central point of the rectangular region custom-character(k1) on the endoscopic image I(k1), where

    [00002] ( k - 1 ) = { p u 1 + u , v 1 + v c ( k - 1 ) .Math. "\[LeftBracketingBar]" - L 2 u , v L 2 - 1 } p u 1 + u , v 1 + v c ( k - 1 ) = ( u 1 + u , v 1 + v ) [0058] a rectangular area custom-character(k) of the kth image is:

    [00003] ( k ) = { p u 2 + u , v 2 + v c ( k ) .Math. "\[LeftBracketingBar]" - L 2 u , v L 2 - 1 } p u 2 + u , v 2 + v c ( k ) = ( u 2 + u , v 2 + v )

    [0059] In the embodiment of the present disclosure, when the three-dimensional key point is tracked, the local region is first determined according to a position of the key point and tracked to reduce mistracking of the three-dimensional key point.

    [0060] In step S13, the first local region is mapped to the first local point cloud and the second local region is mapped to the second local point cloud according to the mapping relation between the endoscopic image and the point clouds, the first three-dimensional key point of the first two-dimensional key point on the first local point cloud is determined, and the second three-dimensional key point on the second local point cloud is acquired through the coordinate transformation.

    [0061] In the present step, the three-dimensional key point is initially located. The three-dimensional key point may be initially located in the following two steps.

    [0062] Firstly, a corresponding three-dimensional key point of a two-dimensional key point on the point cloud is determined through a position of the two-dimensional key point according to the mapping relation between the endoscopic image and the point cloud. Secondly, a tissue in the local region may be approximately deemed as a rigid body; and the transformation matrix between the point clouds is solved through the three-dimensional affine transformation, and a three-dimensional key point on a target point cloud is acquired through a coordinate transformation.

    [0063] Correspondingly, S13 specifically includes: [0064] S131, a depth of the endoscopic image is estimated through a neural network to obtain a depth image corresponding to the endoscopic image, space information and color information of each pixel are acquired from the depth image and the endoscope image respectively through reading in rows, and a first local point cloud custom-character(k1) and a second local point cloud custom-character(k) are acquired; [0065] S132, a first three-dimensional key point P(k1)custom-character.sup.3 of the first two-dimensional key point p(k1) on the first local point cloud custom-character(k1) is determined,


    P(k1)=(p(k1)) [0066] where denotes a mapping relation from custom-character(k1) to custom-character(k1), which is recorded as custom-character(k1).fwdarw.custom-character(k1); [0067] S133, in order to acquire a least squares observation, a feature point pair of the local regions custom-character(k1), custom-character(k) is acquired through the optical flow method and recorded as X and Y, respectively, so that X and Y are in a coordinate transformation relation:

    [00004] Y = [ A t ] [ X 1 ] [0068] where Acustom-character.sup.33, tcustom-character.sup.31, =[A t].sup.Tcustom-character.sup.43 denote parameters of a fitting function, is acquirable from the following formula through least squares:


    =([X1].sup.T[X1]).sup.1[X1].sup.TY [0069] and a transformation matrix of an affine transformation of P(k1), P(k) is:

    [00005] T A = [ A t 0 T 1 ] R 4 4 [0070] where 0.sup.T=(0,0,0); and [0071] S134, a three-dimensional affine transformation is performed on the first three-dimensional key point P(k1), where a matrix form is:


    P(k)=T.sub.AP(k1) [0072] and a nearest point is searched for from the second local point cloud custom-character(k) to obtain an initial position of a second three-dimensional key point P(k).

    [0073] In step S14, the second three-dimensional key point is mapped back to the second local region, so as to acquire the second two-dimensional key point from the next image. The two-dimensional coordinates of the tracked key point are acquired by minimizing the preset optimization function in combination with the initial two-dimensional key point, and the corresponding three-dimensional coordinates are finally acquired.

    [0074] In the present step, the three-dimensional key point is precisely located. In the in-vivo environment, the tissue is dynamic, flexible, and highly similar. Therefore, in the embodiment of the present disclosure, the optimization function is constructed through texture information of a key point neighborhood, and the three-dimensional key point is precisely located by minimizing the optimization function.

    [0075] Specifically, firstly, the second three-dimensional key point P(k) is mapped back to the second local region custom-character(k) according to the mapping relation between the endoscopic image and the point cloud, and a second two-dimensional key point p(k) is acquired from the next image.

    [0076] Then, the two-dimensional coordinates of the tracked key point are acquired by minimizing the preset optimization function in combination with the initial two-dimensional key point p.sup.o(k), and the corresponding three-dimensional coordinates are finally acquired.

    [0077] The above optimization function is as follows:


    custom-character=2custom-character.sub.siftcustom-character.sub.optical [0078] where custom-character denotes the optimization function; [0079] custom-character.sub.sift denotes a cosine similarity of a scale invariant feature transform (SIFT) feature vector:

    [00006] J sift ( u , v , k ) = ( p ( k - 1 ) ) T ( p u + u , v + v ( k ) ) .Math. ( p ( k - 1 ) ) .Math. .Math. ( p u + u , v + v ( k ) ) .Math. [0080] where (p(k1)) denotes a feature descriptor of the first two-dimensional key point p(k1) and is a vector, .Math. denotes a norm of the vector, (u, v) denote coordinates of the second two-dimensional key point p(k), p.sub.u+u,v+v(k) denotes a neighborhood point of the second two-dimensional key point p(k), u and v are coordinate offset of the second two-dimensional key point p(k), (p.sub.u+u,v+v(k)) denotes a feature descriptor of the neighborhood point p.sub.u+u,v+v(k) and is a vector, and [0081] custom-character.sub.optical denotes an effect of optical flow information:

    [00007] J optical ( u , v , k ) = a T b .Math. a .Math. .Math. b .Math. [0082] where vectors a, b are defined as in FIG. 2:


    a=(p.sub.u+u,v+v(k)p(k)).sup.T,b=(p.sup.o(k)p(k)).sup.T.

    and {circumflex over (v)} are acquired by traversing and searching for u and v, so as to satisfy the following expression:

    [00008] ( u ^ ( k ) , v ( k ) ) = arg min ( u , v ) ( J ( u , v , k ) )

    [0083] Two-dimensional coordinates of a tracked key point p.sub.u+u,v+v(k) are acquired after ideal offset ((k), {circumflex over (v)}(k)) is obtained, and then corresponding three-dimensional coordinates are finally acquired according to the mapping relation between the endoscopic image and the point cloud.

    [0084] In the embodiment of the present disclosure, after the transformation matrix is acquired through the three-dimensional affine transformation, the three-dimensional key point is precisely located by constructing the optimization function in combination with the texture information and the optical flow information of the endoscopic image. Therefore, the effect of the indistinct in-vivo environment feature on a tracking result is avoided to a certain extent.

    [0085] In conclusion, compared with the prior art, the present embodiment has the beneficial effects as follows: [0086] 1. In the embodiment of the present disclosure, the selected key point is initially tracked through the three-dimensional affine transformation. The three-dimensional key point in the in-vivo environment is precisely and dynamically located and tracked in combination with the texture information and the optical flow information. [0087] 2. In the embodiment of the present disclosure, the doctor labels the two-dimensional key point on the intraoperative image for subsequent display and update of the key point on the three-dimensional point cloud. Accordingly, the information is transmitted intuitively and accurately, and the operation efficiency is improved. [0088] 3. In the embodiment of the present disclosure, when the three-dimensional key point is tracked, the local region is firstly determined according to the position of the key point and tracked to reduce mistracking of the three-dimensional key point. [0089] 4. In the embodiment of the present disclosure, after the transformation matrix is acquired through the three-dimensional affine transformation, the three-dimensional key point is precisely located by constructing the optimization function in combination with the texture information and the optical flow information of the endoscopic image. Therefore, the effect of the indistinct in-vivo environment feature on the tracking result is avoided to a certain extent.

    Embodiment 2

    [0090] As shown in FIG. 3, an embodiment of the present disclosure provides a minimally invasive key trajectory navigation oriented dynamic tracking method for an in-vivo three-dimensional curve. The method includes: [0091] S21, an endoscopic image is read, an operation path curve is acquired from a current image according to the selection of a doctor, and a plurality of first two-dimensional key points through which the operation path curve passes are acquired; [0092] S22, a first local region encompassing the first two-dimensional key point on the current image is tracked, and a second local region is acquired from a next image; [0093] S23, the first local region is mapped to a first local point cloud and the second local region is mapped to a second local point cloud according to a mapping relation between the endoscopic image and the point clouds, a first three-dimensional key point of the first two-dimensional key point on the first local point cloud is determined, and a second three-dimensional key point on the second local point cloud is acquired through a coordinate transformation; [0094] S24, a dimension of the first local point cloud is reduced to obtain a first two-dimensional point cloud, and a second two-dimensional key point of the first three-dimensional key point on the first two-dimensional point cloud is acquired; [0095] a dimension of the second local point cloud is reduced to obtain a second two-dimensional point cloud, and a third two-dimensional key point of the second three-dimensional key point on the second two-dimensional point cloud is acquired; and [0096] two-dimensional coordinates of a tracked key point on the two-dimensional point cloud are acquired by minimizing a preset optimization function according to the second two-dimensional key point and the third two-dimensional key point; and [0097] S25, three-dimensional coordinates of each tracked key point are acquired according to a mapping relation between the point clouds before and after dimension reduction, curve fitting is performed, and a three-dimensional curve is finally obtained by means of tracking.

    [0098] In the embodiment of the present disclosure, the local region is determined through a position of the key point and tracked to reduce the mistracking of the three-dimensional key point. In combination with texture information of the endoscopic image and shape information, the effect of the indistinct in-vivo environment feature is avoided to a certain extent. The three-dimensional key point is precisely located by constructing the optimization function on the point cloud after dimension reduction. Therefore, the inconsistency of a curve shape under different viewing angles is avoided.

    [0099] Each step of the above technical solution will be described in detail below with reference to specific contents:

    [0100] In step S21, the endoscopic image is read, the operation path curve is acquired from the current image according to selection of the doctor, and the plurality of first two-dimensional key points through which the operation path curve passes are acquired.

    [0101] A first two-dimensional key point acquisition process in S21 includes:


    custom-character.sup.p={p.sub.0.sup.p, . . . ,p.sub.j.sup.p . . . ,p.sub.l1.sup.p} [0102] is defined as a pixel through which the operation path curve passes; where [0103] for a point on the curve, a curvature of a jth pixel on the curve is:

    [00009] K j = v j + - v j u j + - u j , j = 0 , , 2 .Math. , j < l - 1 , [0104] where p.sub.j+.sup.p=[u.sub.j+ v.sub.j+a] denotes coordinates of a j+th pixel on the curve, and denotes an interval number of the pixels when the curvature of the pixel is solved; and [0105] for curvatures of two consecutive pixels, when |K.sub.j+K.sub.j|, denotes a curvature threshold; and all first two-dimensional key points custom-character={p.sub.1, . . . , p.sub.i, . . . , p.sub.n} are determined in combination with p.sub.j+.sup.p as a key point on the operation path curve and a start point and an end point of the curve, where n denotes a total number of the first two-dimensional key points.

    [0106] In the present step, the doctor labels a three-dimensional curve on an intraoperative image for subsequent display and update of the curve on a three-dimensional point cloud. Accordingly, information is transmitted intuitively and accurately, and an operation efficiency is improved. Moreover, the key point is determined according to the curvature of the curve, and a computation speed is ensured for key point tracking.

    [0107] Step S22 that a first local region encompassing the first two-dimensional key point on the current image is tracked, and a second local region is acquired from a next image specifically includes: [0108] S221, firstly, a k1th image is defined as I(k1)custom-character.sup.WH3, where W denotes a width of the endoscopic image, H denotes a height of the endoscopic image, I(k) denotes a kth image, and custom-character(k1)={p.sub.1(k1), . . . , p.sub.n(k1)} denote a plurality of first two-dimensional key points; and [0109] maximums and minimums of all the first two-dimensional key points on a u axis and a v axis of an image coordinate system are determined, respectively, a position of a median on each axis is selected as p.sup.c(k1)=(u.sub.1, v.sub.1)R.sup.2, and a first local region custom-character(k1) is determined according to a preset region shape and side length; [0110] S222, feature matching is performed on feature points of images I(k1), I(k) through the optical flow, and a center p.sup.c(k), corresponding to p.sup.c(k1), of the second local region is acquired from the image I(k),

    [00010] p c ( k ) = p c ( k - 1 ) + 1 m .Math. i = 1 m ( p i f ( k ) - p i f ( k - 1 ) ) [0111] where p.sup.f(k)={p.sub.1.sup.f(k), p.sub.2.sup.f(k), . . . , p.sub.m.sup.f(k)} denotes a feature point on the image I(k), and m denotes the number of the feature point on the image I(k); and [0112] S223, a second local region custom-character(k) is determined according to the center p.sup.c(k) and the preset region shape and side length.

    [0113] In the embodiment of the present disclosure, the local region is determined through a position of the key point and tracked to reduce mistracking of the three-dimensional key point.

    [0114] In step S23, the first local region is mapped to the first local point cloud and the second local region is mapped to the second local point cloud according to the mapping relation between the endoscopic image and the point clouds, the first three-dimensional key point of the first two-dimensional key point on the first local point cloud is determined, and the second three-dimensional key point on the second local point cloud is acquired through the coordinate transformation.

    [0115] In the present step, the three-dimensional key point is initially located. The three-dimensional key point may be initially located in the following two steps.

    [0116] Firstly, a corresponding three-dimensional key point of a two-dimensional key point on the point cloud is determined through a position of the two-dimensional key point according to the mapping relation between the endoscopic image and the point cloud. Secondly, a tissue in the local region may be approximately deemed as a rigid body; and a transformation matrix between the point clouds is solved through a three-dimensional affine transformation, and a three-dimensional key point on a target point cloud is acquired through a coordinate transformation.

    [0117] Correspondingly, S23 specifically includes: [0118] S231, a depth of the endoscopic image is estimated to obtain a depth image corresponding to the endoscopic image, space information and color information of each pixel are acquired from the depth image and the endoscopic image respectively through reading in rows, and a first local point cloud custom-character(k1) and a second local point cloud custom-character(k) are acquired; [0119] S232, a first three-dimensional key point custom-character(k1)={P.sub.1(k1), . . . , P.sub.n(k1)}custom-character(k1), P.sub.i(k1)custom-character.sup.3 of the first two-dimensional key point custom-character(k1) on the first local point cloud custom-character(k1) is determined,


    P.sub.i(k1)=(p.sub.i(k1)) [0120] where denotes a mapping relation from custom-character(k1) to custom-character(k1), which is recorded as custom-character(k1).fwdarw.custom-character(k1); [0121] S233, a feature point pair of the local regions custom-character(k1), custom-character(k) is acquired through the optical flow method and recorded as X and Y, respectively, so that X and Y are in a coordinate transformation relation:

    [00011] Y = [ A t ] [ X 1 ]

    where Acustom-character.sup.33, tcustom-character.sup.31, =[A t].sup.Tcustom-character.sup.43 denote parameters of a fitting function, is acquirable from the following formula through least squares:


    =([X1].sup.T[X1]).sup.1[X1].sup.TY [0122] and a transformation matrix of an affine transformation of P(k1), P(k) is:

    [00012] T A = [ A t 0 T 1 ] 4 4 [0123] where 0.sup.T=(0,0,0); and [0124] S234, a three-dimensional affine transformation is performed on the first three-dimensional key point custom-character(k1), and a nearest point is searched for from the second local point cloud custom-character(k) to obtain an initial position of a second three-dimensional key point custom-character(k), where


    custom-character(k)=T.sub.Acustom-character(k1).

    [0125] In step S24, the dimension of the first local point cloud is reduced to obtain the first two-dimensional point cloud, and the second two-dimensional key point of the first three-dimensional key point on the first two-dimensional point cloud is acquired; [0126] the dimension of the second local point cloud is reduced to obtain the second two-dimensional point cloud, and the third two-dimensional key point of the second three-dimensional key point on the second two-dimensional point cloud is acquired; and [0127] the two-dimensional coordinates of the tracked key point on the two-dimensional point cloud are acquired by minimizing the preset optimization function according to the second two-dimensional key point and the third two-dimensional key point.

    [0128] In the present step, the three-dimensional key point is precisely located. In order to ensure the accuracy when the curve dynamically changes, it is required to optimize the initial position of the key point on the curve.

    [0129] In the present step, firstly, in combination with the texture information of the endoscopic image and the shape information, the effect of the indistinct in-vivo environment feature is avoided to a certain extent. Secondly, the three-dimensional key point is precisely located by constructing the optimization function on the point cloud after dimension reduction. Therefore, the inconsistency of a curve shape under different viewing angles is avoided.

    [0130] Specifically, firstly, a dimension of the first local point cloud custom-character(k1) is reduced to obtain a first two-dimensional point cloud Q(k1), and an ith second two-dimensional key point T.sub.i(k1) of the first three-dimensional key point custom-character(k1) on the first two-dimensional point cloud Q(k1) is acquired; and [0131] a dimension of the second local point cloud custom-character(k) is reduced to obtain a second two-dimensional point cloud Q(k), and an ith third two-dimensional key point T.sub.i(k) of the second three-dimensional key point custom-character(k) on the second two-dimensional point cloud Q(k) is acquired.

    [0132] Then two-dimensional coordinates of the tracked key point on the two-dimensional point cloud are acquired by minimizing the preset optimization function according to the second two-dimensional key point and the third two-dimensional key point.

    [0133] The optimization function is as follows:


    custom-character=1custom-character.sub.sift+custom-character.sub.shape [0134] where custom-character denotes the optimization function; [0135] custom-character.sub.sift denotes a cosine similarity of an SIFT feature vector:

    [00013] J sift ( T i ( k ) ) = ( T i ( k - 1 ) ) T ( T i ( k ) ) .Math. ( T i ( k - 1 ) ) .Math. .Math. ( T i ( k ) ) .Math. [0136] where (T.sub.i(k1)) denotes a feature descriptor of the i th second two-dimensional key point T.sub.i(k1) and is a vector, (T.sub.i(k)) denotes a feature descriptor of a neighborhood point of the ith third two-dimensional key point T.sub.i(k) and is a vector, and .Math. denotes a norm of the vector; and [0137] custom-character.sub.shape denotes a difference in cosine values of included angles between adjacent key points on different curves:

    [00014] shape ( T i ( k ) ) = 1 n .Math. i = 0 n .Math. "\[LeftBracketingBar]" ( T i ( k - 1 ) ) - ( T i ( k ) ) .Math. "\[RightBracketingBar]" [0138] where g(T.sub.i(k)) is the cosine value of the included angle, which is specifically calculated as follows:

    [00015] ( T i ( k ) ) = a T b .Math. a .Math. .Math. b .Math. where a = ( T i + 1 ( k ) - T i ( k ) ) T , b = ( T i - 1 ( k ) - T i ( k ) ) T .

    custom-character is minimized by traversing and searching for neighborhood points of all third two-dimensional key points T.sub.i(k), so as to satisfy:


    {circumflex over (T)}.sub.i=argmin(1J.sub.sift(T.sub.i(k))+custom-character.sub.shape(T.sub.i(k)))

    [0139] An ideal key point T.sub.i set can be acquired by minimizing the optimization function.

    [0140] In the embodiment of the present disclosure, after the transformation matrix is acquired through the three-dimensional affine transformation, the three-dimensional key point is precisely located by constructing the optimization function in combination with the texture information of the endoscopic image and the shape information. Therefore, the effect of the indistinct in-vivo environment feature on a tracking result is avoided to a certain extent.

    [0141] In step S25, the three-dimensional coordinates of each tracked key point are acquired according to the mapping relation between the point clouds before and after dimension reduction, curve fitting is performed, and the three-dimensional curve is finally obtained by means of tracking.

    [0142] Specifically, in the present step, interpolation fitting is performed on a line through an equation of a B-spline curve, where a general equation of the B-spline curve is:


    P(t)=.sub.i=0.sup.mP.sub.iF.sub.i,k(t) [0143] where P.sub.i denotes a feature point of a control curve, F.sub.i,k(t) denotes a kth order B-spline basis function, and the three-dimensional curve is tracked through curve interpolation fitting. P.sub.i is acquired by performing mapping on T.sub.i according to the mapping relation between the point clouds before and after dimension reduction.

    [0144] In conclusion, compared with the prior art, the present embodiment has the beneficial effects as follows: [0145] 1. In the embodiment of the present disclosure, the local region is determined through a position of the key point and then tracked to reduce mistracking of the three-dimensional key point. In combination with the texture information of the endoscopic image and the shape information, the effect of the indistinct in-vivo environment feature is avoided to a certain extent. The three-dimensional key point is precisely located by constructing the optimization function on the point cloud after dimension reduction. Therefore, the inconsistency of a curve shape under different viewing angles is avoided. [0146] 2. In the present step, the doctor labels the two-dimensional curve on the intraoperative image for subsequent display and update of the curve on the three-dimensional point cloud. Accordingly, the information is transmitted intuitively and accurately, and an operation efficiency is improved; and moreover, the key point is determined according to the curvature of the curve, and a computation speed is ensured for key point tracking. [0147] 3. In the embodiment of the present disclosure, after the transformation matrix is acquired through the three-dimensional affine transformation, the three-dimensional key point is precisely located by constructing the optimization function in combination with the texture information of the endoscopic image and the shape information. Therefore, the effect of the indistinct in-vivo environment feature on the tracking result is avoided to a certain extent.

    Embodiment 3

    [0148] An embodiment of the present disclosure provides a storage medium. The storage medium stores an autonomous operation robot oriented computer program for three-dimensional key point tracking, where the computer program causes a computer to execute the above tracking method for an in-vivo three-dimensional key point.

    Embodiment 4

    [0149] An embodiment of the present disclosure provides an electronic apparatus. The electronic apparatus includes: [0150] one or more processors; [0151] a memory; and [0152] one or more programs, where the one or more programs are stored in the memory and configured to be executed by the one or more processors, and the program includes executing the above tracking method for an in-vivo three-dimensional key point.

    [0153] It can be understood that the storage medium and the electronic apparatus according to Embodiment 3 and Embodiment 4 of the present disclosure respectively correspond to the minimally invasive key site navigation oriented dynamic tracking method for an in-vivo three-dimensional key point according to Embodiment 1 of the present disclosure. Reference may be made to the corresponding parts of the tracking method for a three-dimensional key point for the explanations, instances, beneficial effects, etc. of the relevant contents of the storage medium and the electronic apparatus, which will not be repeated herein.

    [0154] In conclusion, in all the above embodiments, the three-dimensional key point in the in-vivo environment can be precisely and dynamically located and tracked. In addition, for dynamic tracking of the in-vivo three-dimensional curve, since the three-dimensional key point can be precisely and dynamically located and tracked, the in-vivo three-dimensional curve can also be precisely and dynamically located and tracked while the operation path curve can be tracked in real time.

    [0155] It is to be noted that relational terms herein such as first and second are merely used to distinguish one entity or operation from another entity or operation without necessarily requiring or implying any such an actual relation or order between these entities or operations. In addition, terms comprise, include, encompass, or any other their variations are intended to cover a non-exclusive inclusion. Therefore, a process, method, article, or apparatus including a series of elements not only includes those elements, but also includes other elements that are not explicitly listed, or further includes inherent elements of such a process, method, article, or apparatus. Without more restrictions, the elements defined by the sentence comprise a . . . and include a . . . do not exclude the existence of other identical elements in the process, method, article, or apparatus including the elements.

    [0156] The above embodiments are only used to explain the technical solutions of the present disclosure, and are not intended to limit same. Although the present disclosure is described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that they can still made modifications to the technical solutions described in all the foregoing embodiments, or make equivalent substitutions to some technical features in the embodiments. These modifications or substitutions do not make the essence of the corresponding technical solutions depart from the spirit and scope of the technical solutions in all the embodiments of the present disclosure.