SYSTEM AND METHOD FOR RETINA TEMPLATE MATCHING IN TELEOPHTHALMOLOGY
20220015629 · 2022-01-20
Assignee
Inventors
- Eric J. Seibel (Seattle, WA)
- Chen GONG (Seattle, WA, US)
- Steven L. BRUNTON (Seattle, WA, US)
- Nils Benjamin ERICHSON (Seattle, WA, US)
- Laura TRUTOIU (Seattle, WA, US)
- Brian T. Schowengerdt (Seattle, WA, US)
Cpc classification
A61B2576/02
HUMAN NECESSITIES
G06V10/76
PHYSICS
A61B3/0025
HUMAN NECESSITIES
G06T2207/10101
PHYSICS
A61B3/12
HUMAN NECESSITIES
A61B3/14
HUMAN NECESSITIES
International classification
A61B3/12
HUMAN NECESSITIES
A61B3/00
HUMAN NECESSITIES
A61B3/14
HUMAN NECESSITIES
Abstract
A retina image template matching method is based on the registration and comparison between the images captured with portable low-cost fundus cameras (e.g., a consumer grade camera typically incorporated into a smartphone or tablet computer) and a baseline image. The method solves the challenges posed by registering small and low-quality retinal template images captured with such cameras. Our method combines dimension reduction methods with a mutual information (MI) based image registration technique. In particular, principle components analysis (PCA) and optionally block PCA are used as a dimension reduction method to localize the template image coarsely to the baseline image, then the resulting displacement parameters are used to initialize the MI metric optimization for registration of the template image with the closest region of the baseline image.
Claims
1-38. (canceled)
39. A method for monitoring a retina of a subject, comprising the steps of: obtaining a set of retina template images of a retina; matching a retina template image in the set to a retina baseline image of the retina at least by: applying dimension reduction to the retina baseline image and the retina template image; and generating a registered retina template image at least by registering the retina template image to at least a portion of the retina baseline image, wherein the retina baseline image captures a wider field of view than the retina template image; and registering the retina template image as a registered retina template image at least based at least in part upon at least a portion of the retina baseline image, wherein the retina baseline image comprises a wider field of view than the retina template image; and detecting a change in a characteristic pertaining to the retina or a subject of the retina based at least in part upon a comparison between the registered retina template image and the retina baseline image.
40. The method of claim 39, matching the retina template image to the retina baseline image comprising: generating a plurality of smaller offset target images at least by cropping the retina baseline image, wherein at least two of the plurality of smaller offset target images overlap each other.
41. The method of claim 40, matching the retina template image to the retina baseline image comprising: mapping a smaller offset target image of the plurality of smaller offset target images into a lower-dimensional space based at least in part upon a first result of a principal component analysis performed on the plurality of smaller offset target images, where the lower-dimensional space has a smaller dimensionality than a dimensional space of the retina baseline image.
42. The method of claim 41, matching the retina template image to the retina baseline image further comprising: performing coarse localization on the retina template image or the set of retina template images based at least in part upon the principal component analysis performed on the retina template image based at least in part upon a first result of the principal component analysis.
43. The method of claim 42, matching the retina template image to the retina baseline image further comprising: determining a nearer target image based at least in part upon a result of performing the coarse localization to the retina template image or the set of retina template images.
44. The method of claim 43, matching the retina template image to the retina baseline image further comprising: determining a displacement parameter in a domain of a mutual information metric for registration of the retina template image with the nearer target image based at least in part upon the principal component analysis performed on the retina template image.
45. The method of claim 44, matching the retina template image to the retina baseline image further comprising: initializing metric processing on the mutual information metric at least by using at least the displacement parameter; and registering the retina template image with the nearer target image based at least in part upon mutual information registration and a location of the retina template image onto the retina baseline image.
46. The method of claim 45, matching the retina template image to the retina baseline image further comprising: after the retina template image is registered with the nearest target image, locating the retina template image onto the retina baseline image.
47. The method of claim 39, further comprising updating the retina baseline image with a mosaic image having the set of retina template images registered to the retina baseline image, wherein the imaging capturing device comprises a head-worn retinal imaging device, a fundus camera which further includes an optical coherence tomography (OCT) component, a portable scanning laser ophthalmoscope, a hand-held device including a camera adapted for imaging the retina, or a camera embodied in a smartphone or tablet computer configured with apparatus to assist in taking an image of the retina.
48. The method of claim 39, wherein the change in the characteristic of the retina or the subject comprises hypertension, abusive head trauma, diabetic retinopathy, or glaucoma.
49. The method of claim 39, further comprising: mosaicking the retina template image into a new retina baseline image; determining a separate set of retina template images; comparing the separate set of retina template images with the new retina baseline image to generate a comparison result; and determining whether the characteristic pertaining to the retina or the subject has degraded based at least in part upon the comparison result.
50. The method of claim 49, mosaicking the retina template image comprising: determining a set of smaller images for the retina template image; and mapping the set of smaller images into a mapped set of smaller images in a lower-dimensional space based at least in part upon a principal component analysis, wherein the lower-dimensional space has a lower dimensionality than a separate dimensional space of the retina template image.
51. The method of claim 50, mosaicking the retina template image comprising: generating a set of lower-dimensional features in the lower-dimensional space for the retina template based at least in part upon the mapped set of smaller images; determining a distance metric between two lower-dimensional features in the set of lower-dimensional features in the lower-dimensional space; determining, for a smaller image of the set of smaller images, one or more closer target images in the lower-dimensional space based at least in part upon an extent of overlap between a pair of the smaller image and a closer target images of the one or more closer target images; and determining a mutual information metric for the pair of the smaller image and the closest target image.
52. The method of claim 51, mosaicking the retina template image comprising: registering, for the smaller image, the one or more closest target images based at least in part upon the mutual information metric; and generating a stitched image at least by aligning the one or more closest target images with at least the mutual information metric.
53. An extended reality device, comprising: an imaging device; a processor operatively coupled to the imaging device; and a memory storing therein a sequence of instructions which, when executed by the processor, causes the processor to perform a set of acts, the set of acts comprising: obtaining a set of retina template images of a retina; matching a retina template image in the set to a retina baseline image of the retina at least by: applying dimension reduction to the retina baseline image and the retina template image; and generating a registered retina template image at least by registering the retina template image to at least a portion of the retina baseline image, wherein the retina baseline image captures a wider field of view than the retina template image; and registering the retina template image as a registered retina template image at least based at least in part upon at least a portion of the retina baseline image, wherein the retina baseline image comprises a wider field of view than the retina template image; and detecting a change in a characteristic pertaining to the retina or a subject of the retina based at least in part upon a comparison between the registered retina template image and the retina baseline image.
54. The extended reality device of claim 53, wherein the set of acts further comprising: generating a plurality of smaller offset target images at least by cropping the retina baseline image, wherein at least two of the plurality of smaller offset target images overlap each other; mapping a smaller offset target image of the plurality of smaller offset target images into a lower-dimensional space based at least in part upon a first result of a principal component analysis performed on the plurality of smaller offset target images, where the lower-dimensional space has a smaller dimensionality than a dimensional space of the retina baseline image; and performing coarse localization on the retina template image or the set of retina template images based at least in part upon the principal component analysis performed on the retina template image based at least in part upon a first result of the principal component analysis.
55. The extended reality device of claim 53, wherein the set of acts further comprising: mosaicking the retina template image into a new retina baseline image; determining a separate set of retina template images; comparing the separate set of retina template images with the new retina baseline image to generate a comparison result; determining whether the characteristic pertaining to the retina or the subject has degraded based at least in part upon the comparison result; determining a set of smaller images for the retina template image; and mapping the set of smaller images into a mapped set of smaller images in a lower-dimensional space based at least in part upon a principal component analysis, wherein the lower-dimensional space has a lower dimensionality than a separate dimensional space of the retina template image.
56. A non-transitory machine accessible storage medium having stored thereupon a sequence of instructions which, when executed by a processor of a mixed reality device, causes the processor to perform a set of acts, the set of acts comprising: obtaining a set of retina template images of a retina; matching a retina template image in the set to a retina baseline image of the retina at least by: applying dimension reduction to the retina baseline image and the retina template image; and generating a registered retina template image at least by registering the retina template image to at least a portion of the retina baseline image, wherein the retina baseline image captures a wider field of view than the retina template image; and registering the retina template image as a registered retina template image at least based at least in part upon at least a portion of the retina baseline image, wherein the retina baseline image comprises a wider field of view than the retina template image; and detecting a change in a characteristic pertaining to the retina or a subject of the retina based at least in part upon a comparison between the registered retina template image and the retina baseline image.
57. The non-transitory machine accessible storage medium of claim 56, wherein the set of acts further comprising: generating a plurality of smaller offset target images at least by cropping the retina baseline image, wherein at least two of the plurality of smaller offset target images overlap each other; mapping a smaller offset target image of the plurality of smaller offset target images into a lower-dimensional space based at least in part upon a first result of a principal component analysis performed on the plurality of smaller offset target images, where the lower-dimensional space has a smaller dimensionality than a dimensional space of the retina baseline image; and performing coarse localization on the retina template image or the set of retina template images based at least in part upon the principal component analysis performed on the retina template image based at least in part upon a first result of the principal component analysis.
58. The non-transitory machine accessible storage medium of claim 56, wherein the set of acts further comprising: mosaicking the retina template image into a new retina baseline image; determining a separate set of retina template images; comparing the separate set of retina template images with the new retina baseline image to generate a comparison result; determining whether the characteristic pertaining to the retina or the subject has degraded based at least in part upon the comparison result; determining a set of smaller images for the retina template image; and mapping the set of smaller images into a mapped set of smaller images in a lower-dimensional space based at least in part upon a principal component analysis, wherein the lower-dimensional space has a lower dimensionality than a separate dimensional space of the retina template image.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0030] The appended drawing figures are offered by way of example and not limitation of currently preferred embodiments of this disclosure.
[0031]
[0032]
[0033] (a) create a multitude of offset target images from a full/baseline image;
[0034] (b) create low-dimensional representations of each of the target images using PCA;
[0035] (c) perform coarse localization of the template image: find the nearest target image in the low-dimensional space; and
[0036] (d) MI-based registration of the template and the nearest target image, and locate the template image onto the baseline image.
[0037]
[0038]
[0039]
[0040]
[0041]
DETAILED DESCRIPTION
[0042] This document discloses an efficient and accurate retinal matching system and method combining dimension reduction and mutual information (MI), we refer to the technique here as RetinaMatch. By way of overview, the dimension reduction initializes the MI optimization as a coarse localization process, which narrows the optimization domain and avoids local optima. The disclosed system and method outperforms to-date existing template matching solutions. In addition, we disclose a system and method for image mosaicking with area-based registration, providing a robust approach which may be used when the feature-based methods fail. To the best of our knowledge, this is the first template matching technique for retina images with small template images from unconstrained retinal areas.
[0043] Our approach is an improvement over the area-based matching methods with MI metric, since it is more reliable to achieve accurate and robust template matching near the alignment position. One unique aspect of our approach is that we combine dimension reduction methods with the MI-based registration to reduce the interference of local extrema and improve the matching efficiency.
[0044] An example of the practical use of our method in monitoring the retina in a teleophthalmology setting is shown in
[0045] It is also contemplated that the template images captured by the smartphone 12 could be sent over the network 16, 18 to a computing system 22 in the eye clinic and the processing steps of
[0046] Specific examples of the applications of the retinal template matching in a teleophthalmology setting will be discussed at length later in this document.
[0047] With the above description in mind, one of the principal aspects of this disclosure is application of dimension reduction methods with the MI-based registration to reduce the inference of local extrema and improve the matching efficiency.
[0048]
[0049] The process of
[0050] The procedures shown in panels (a) and (b) of
[0051] With the above explanation in mind, attention will now be directed to
[0052]
[0053] A specific embodiment of our procedure of
1.
[0054] We define the full image and template as F and S respectively. The full image F is split into target images I.sub.1, I.sub.2, . . . , I.sub.N:
I.sub.i=Ø(b.sub.i,F).
[0055] The function Ø crops the target images I.sub.i from F at b, and b.sub.i=[x.sub.i, y.sub.i, h, w], where (x.sub.i,y.sub.i) denotes the center position and (h,w) denotes the width and height of the cropped image. There is a certain displacement, f of neighboring target images in the x and y axes. As shown in
[0056] Target images are resized to vectors and form the matrix X∈.sup.n×d.
2.
[0057] Dimension reduction methods allow the construction of low-dimensional summaries, while eliminating redundancies and noise in the data. To estimate the template location in the 2d space, the full image dimension is redundancy, thus we apply dimension reduction methods for the template coarse localization.
[0058] Generally, we can distinguish between linear and nonlinear dimension reduction techniques. The most prominent linear technique PCA. PCA is selected as the dimension reduction method in RetinaMatch since PCA is simple and versatile. Specifically, PCA forms a set of new variables as a weighted linear combination of the input variables. Consider a matrix X=[x.sub.1, x.sub.2, . . . , x.sub.d] of dimension n×d, where n denotes the number of observations and d is the number of variables. Further, we assume that the matrix X is column-wise mean centered. The idea of PCA is to form a set of uncorrelated new variables (referred to as principal components) as a linear combination of the input variables:
z.sub.i=Xw.sub.i, (1)
where z.sub.i is the .sub.ith principal component (PC) and w.sub.i is the weight vector. The first PC explains most of the variation in the data, the subsequent PCs then account for the remaining variation in descending order. Thereby, PCA imposes the constraint that the weight vectors are orthogonal. This problem can be expressed compactly as the following minimization:
minimize ∥X−ZW∥.sub.F.sup.2
subject to W.sup.TW=I
where ∥.∥.sub.F is the Frobenius norm. The weight matrix W that maps the input data to a subspace turns out to be the right singular vectors of the input matrix X. Often a low-rank approximation is desirable, e.g., we compute the k dominant PCs using a truncated weight matrix W.sub.k=[w.sub.1, w.sub.2, . . . , w.sub.k]. k is some integer, such as 20.
[0059] PCA is generally computed by the singular value decomposition (SVD). Many algorithms have been developed to streamline the computation of the SVD or PCA for high-dimensional data that exhibits low-dimensional patterns, see J. N. Kutz, et al., Dynamic mode decomposition: data-driven modeling of complex systems. SIAM, 2016, vol. 149. In particular, tremendous strides have been made accelerating the SVD and related computations using randomized methods for linear algebra. See the references 24-31 cited in the manuscript portion of the priority U.S. provisional application. Since we have demonstrated high performance with less than 20 principal components, the randomized SVD is used to compute the principal components, improving the efficiency in this retinal mapping application for mobile device platforms (e.g., smartphone, tablet). The randomized algorithm proceeds by forming a sketch Y of the input matrix
Y=XΩ,
Where Ω is a d×l random test matrix, say with independent and identically distributed random standard normal entries. Thus, the l columns of Y are formed as a randomly weighted linear combination of the columns of the input matric, providing a basis for the column space of X. Note, that l is chosen to be slightly larger than the desired number of principal components. Next, we form an orthonormal basis Q using the QR decomposition Y=QR. Now, we use this basis matrix to project the input data matrix to low-dimensional space
B=Q.sup.TX.
This smaller matrix B of dimension l×d can then be used to efficiently compute the low-rank SVD and subsequently the dominant principal components. Given the SVD of B=UΣV.sup.T, we obtain the approximate principal components as
Z=QUΣ=XV.
[0060] Here, U and V are the left and right singular vectors and the diagonal elements of Σ are the corresponding singular values. The approximation accuracy can be controlled via additional oversampling and power iterations.
[0061] Referring again to panel (b) of
Z=XW,
Where Z=[z.sub.1, z.sub.2, z.sub.3, . . . , z.sub.N].sup.T∈, w∈
.sup.d×l and l<<d. The image space Ω.sub.1 is mapped to a low-dimensional space Ω.sub.2 with the mapping W. W and Z are saved in memory, in what we have called a “dictionary”, D.
[0062] It is important to note that PCA is sensitive to outliers, occlusions, and corruption in the data. In ophthalmological imaging applications, there are several potential sources of corruption and outliers when imaging the full image, including blur, uncorrected astigmatism, inhomogeneous illumination, glare from crystalline lens opacity, internal reflections (e.g., from the vitreoretinal interface and lens), transient floaters in the vitreous, and shot noise in the camera. Further, there is often a trade-off between illumination and image quality, and there is strong motivation to introduce as little light as necessary for the patient comfort and health. The robust principal component analysis (RPCA) was introduced specifically to address this issue, decomposing a data matrix into the sum of a matrix containing low-rank coherent structure and a sparse matrix of outliers and corrupt entries. In general, RPCA is more expensive than PCA, requiring an iterative optimization to decompose the original matrix into sparse and low-rank components. Each step of the iteration is as expensive as regular PCA, and typically on the order of tens of iterations are required; however, PCA may be viewed as an offline step in our procedure, so that this additional computational cost is manageable. RPCA has been applied with success in retinal imaging applications to improve image quality. In the examples presented in this work, the data appears to have few enough outliers so that RPCA is not necessary, although it is important to keep RCPA as an option for data with outliers and corruption. Further details on RPCA are contained in the references cited in the manuscript portion of our prior provisional application.
3. FIG. 2 Panel (c): Coarse Localization—Find the Nearest Target Image in the Low-Dimensional Space
[0063] Given a template S, the coarse position can be estimated by recognizing its nearest target image. The nearest target image in the image space Ω.sub.1 should also be the nearest representation of S in the lower dimensional space Ω.sub.2. Accordingly, we obtain the low-dimension feature z.sub.s of the template in Ω.sub.2:
z.sub.s=.sub.{tilde over (S)}W,
where s∈.sup.d is the reshaped vector of template S. Let Δ (z.sub.s, z) be the Euclidean distance between z.sub.s and a feature z in Z. z* is the nearest target feature of the source image S in Ω.sub.2:
z*=arg.sub.zmin Δ(z.sub.s,z).
[0064] The corresponding target image location is used as the coarse location of S. Ideally, the difference between the coarse location and the ground truth in x and y axes should be less than f/2 pixels.
[0065] In one of the experiments we performed, PCA outperforms other non-linear dimension reduction methods, while the error is larger than f/2. The main reason is that the image degradation creates spurious features that contribute to the final classification. To reduce the influence of local features, we implement block PCA to further improve the accuracy of the coarse localization. By computing the PCA of different local patches in the template, the effect of local features, which cannot be located correctly, is reduced. This procedure is shown in
[0066] Obtaining the nearest target image, we crop a larger image at the same position from the full image as the new target image I. In this way, the template can have more overlap with the new target image when there is a large offset between two images. We segment I and the template S into small patches with the cropping function .sub.{tilde over (Ø)}, where the patch size is smaller than the source image with the axial displacement of neighboring patches f′ Similarly, all image patches from I are mapped into the low-dimensional space Ω.sub.3 with W Let Z′ denote the low-dimensional representation of the target image distribution. Each template patch is then mapped to the space with W′. The nearest target patch for each template patch is determined with the Euclidean distance as described before. We use the same weight for each region of the template for localization. Let b.sub.m be the mean of the coordinates of selected nearest target patches, which then represents the center of the template on I. Accordingly, the template location on the full image can be estimated and the region is cropped as the image Ŝ. We store the representation of each of the target image patches in lower dimensional space in memory, referred to as “dictionary” T. The accurate registration is then applied to the template S and image Ŝ. In this way the coarse localization provides an estimate of a good initial point for the accurate registration (panel (d) of
[0067] In the implementation of the proposed coarse localization, the full (baseline) image is assumed to exist so the dictionary D and dictionary T for each target image can be built in advance. This is the pre-computed part as shown in
[0068]
[0069] Example processing instructions for coarse localization:
TABLE-US-00001 1 Map template S into space Ω.sub.2: z.sub.s = {tilde over (s)}W. 2 Determine closest target image I with corresponding z*: z* = arg min.sub.z Δ(z.sub.s, z), z* ∈ Z. 3 Segment S into [S.sub.p.sup.1, S.sub.p.sup.2, . . . , S.sub.p.sup.n]: S.sub.p.sup.i = {tilde over (ϕ)}(b.sub.i, S); Segment I into [I.sub.p.sup.1, I.sub.p.sup.2, . . . , I.sub.p.sup.n]: I.sub.p.sup.i = {tilde over (ϕ)}(b.sub.i, I). 4 Map target patches I.sub.p.sup.i into space Ω.sub.3: Z′ = I.sub.p W′, where I.sub.p is formed with vectorized I.sub.p.sup.i. 5 For each template patch S.sub.p.sup.i: 6 (i)Map S.sub.p.sup.i into space Ω.sub.3: {tilde over (z)}.sub.s.sup.i =S.sub.p.sup.i W′. 7 (ii)Determine its closest target patch I.sub.p.sup.Idx(i) with index Idx(i). 8
4.
[0070] Panel (d) of
[0071] (1) Image Registration Between Template and Nearest Target Image Using Mutual Information (MI) (
[0072] In this section, we describe the maximization of MI for multimodal image registration. We define images S and Ŝ as the template and target images, respectively. A transform u is defined to map pixel locations x∈S to pixel locations in Ŝ.
[0073] The main idea of the registration is to find a deformation û at each pixel location x that maximizes the MI between the deformed template image S(u(x)) and the target image Ŝ(x). Accordingly,
Where
[0074]
Here, i.sub.1 and i.sub.2 are the image intensity values in S(u(x)) and Ŝ(x), respectively and p(i.sub.1) and p(i.sub.2) are their marginal probability distributions while p(i.sub.1, i.sub.2) is their joint probability distribution. The probability distributions p(i.sub.1, i.sub.2) reflect the degree to which the greyscale (image intensity) values of each pixel in S(u(x)) and Ŝ(x) are similar; p(i.sub.1, i.sub.2) has a high value (closer to 1) if the pixel values are similar, and low value (closer to 0) if the pixel values are dissimilar. In more detail, in terms of mutual information, based on discrete data like images, each pixel has a grayscale value for 0 to 255. (Although examples herein may describe use of grayscale images for the fundus image work, embodiments are not so limited and may also employ color images as appropriate). We first compute the joint histogram of two images: the joint histogram will be 256×256, which counts the number of corresponding pixels' grayscale from two images. For example, if in the first pixel, one image has a grayscale of 100 and another one is 120, then the joint histogram map (100, 120) will add one. After we finished the joint histogram, the joint probability p(i.sub.1, i.sub.2) can be obtained by normalizing the joint histogram. Then the marginal probability is computed according to:
[0075] (2) Locate the Template onto the Full Image (
[0076] In this step images S and Ŝ are accurately registered with maximization of mutual information, as per sub-step (d)(1) above. The location of mage Ŝ on the full image F becomes the estimated displacement of the template S. In our work, the transform u for alignment is given as an affine transformation:
[0077] It will be appreciated that the processing to create the target images and map them into lower dimensional space (panels (a) and (b) of
[0078] Image Mosaicking
[0079]
[0080] As pointed out previously, the full retina image can be stitched into a panorama by using many small templates. Users must capture a series of images in naturally unconstrained eye positions to explore different regions of the retina. It is problematic to determine adjacent images before the registration when we apply area-based registration approaches, because at that time they may not have effective descriptors for matching.
[0081] Related to the dimension reduction in the proposed template matching method, here we present the procedure shown in the table below to learn the positional relationship of images to be stitched. In this way, the adjacent images can be recognized and registered efficiently.
[0082] For a series of small images X.sub.i, we form the matrix X. PCA is applied to X and returns the low-dimensional features for each image in Ω.sub.2. The distance between features in Ω.sub.2 indicates the distance between images. We find the nearest N (e.g., N=3) target neighbors in the low dimensional space. The nearest neighbor of image X.sub.i is the one with largest overlap; the image pair is then registered with MI-based approach. To improve the algorithm robustness, the first N nearest neighbors for each image are first selected to compute MI with, and we keep the one with the largest metric value. The above procedure can be represented in the following pseudocode.
[0083] Processing instructions: Image stitching (with reference to
[0091] Applications
[0092] Our method of template matching with baseline images and image mosaicking allows for longitudinal comparisons with previously obtained fundus images of the patient. Such longitudinal comparisons have several applications in the field of ophthalmology as will be described below. Such applications are examples of how the methods of this disclosure can be practiced in a teleophthalmology setting. Other suitable applications are also supported by the embodiments described herein, including options outside of the field of retinal template matching.
[0093] Hypertension
[0094] In the retinal symptom of hypertension, the larger arteries constrict and the venous vessels enlarge in diameter. Ophthalmologists can select several detection points on the vessels. With the captured images coming from the patient as per
[0095] Abusive Head Trauma
[0096] The biomarkers of abusive head trauma (AHT) is another example. The most common retinal manifestation of AHT is multiple retinal hemorrhages in multiple layers of the retina. Matching the captured images onto the full retina image, the hemorrhagic spots can be easily segmented after the subtraction of the current retina regions and previous status. The AHT then can be recognized automatically when such spots are detected. This method permits identification of AHT from images obtained with portable fundus cameras.
[0097] Diabetic Retinopathy
[0098] The obvious symptoms of diabetic retinopathy (DR) are retina hemorrhages and the presence of exudate. They can be monitored follow the similar process of AHT screening.
[0099] Glaucoma
[0100] Glaucoma can cause the optic nerve cup to enlarge. Our matching method can automatically select the images that cover the optic nerve. The following segmentation can be easily implemented and a computation of the optic cup diameter performed. Enlargement of the optic nerve cup over time can be ascertained by comparing the computations from a current image with an image from a previous point in time.
[0101] Use RetinaMatch as a General Image Template Matching Method
[0102] Besides the retina images, the technique of RetinaMatch can be used in other type of image template matching tasks. Note that our method of
[0103] Use RetinaMatch for Camera Localization
[0104] Having the image of the full view, our method of
[0105] Augmented Reality (AR), Eye Glasses, Etc. And Monitoring Changes Over Time
[0106] A retinal imaging system (e.g., consumer grade camera with ancillary imaging device, e.g., D-eye) can be portable and further, can be worn as integrated into, for example, glasses, or an Augmented Reality (AR), Virtual Reality (VR) and/or Mixed Reality (MR) headset, allowing a series of images to be taken and analyzed, either daily, weekly, monthly, or when the user or ophthalmologist requests. These measurements can be discrete, continual, but in a time series and analyzed longitudinally over the increasing time period. Change in a retina can detected by registering and comparing the captured small FOV images to a full baseline retina image using our template matching method.
[0107] AR, VR and/or MR devices can be used to optically scan the retina to form images and thereby acquire the template images. Even more pragmatically, spectacles or sunglasses can be used because of the smaller size, lower costs, and increasing utility to the user. A scanned light beam entering the pupil of the eye and striking the retina to form video rate images perceived by the user's brain can also be used to acquire images of high contrast structures, such as the vasculature containing blood.
[0108] A device can operate without major changes in performance during its lifetime and can be used as a monitor of the condition of a user's eye. By comparing retinal images from such a device over time, the changes in the user's optical system (such as cornea, intraocular lens, retina, and liquid environments) can be monitored to alert the user in possible health changes. For example, these changes can be gradual, like increasing light scattering from the crystalline or intraocular lens due to cataract formation, or the appearance and structural changes in the retina due to diabetic retinopathy. In addition, chronic diseases which may have variations over time in the blood vessel size and shape in conditions of hypertension are another example. Acute changes such as bleeding within the retinal can indicate brain trauma. Relative and repeatable changes in number, size, and shape of structures in the retinal images may indicate that the measured change is due to a particular disease type and not that the AR, VR, MR, glasses, or other type of monitoring device has slowly or suddenly changed its imaging performance or has become unstable.
[0109] However, in many healthy users the optical system will be unchanging over time. In this case, the vasculature of the retina can be used as a test target for detecting optical misalignments, focus errors, light scanning errors and distortions, non-uniform and color imbalance in the illumination, and aberrations in the imaging system. This situation can occur if the monitoring device, such as an AR, VR, or MR device is degraded due to mechanical impact, breakage, applied stresses, applied vibration, thermal changes, and opto-electrical disruption or interference. These changes can be observed in a measurable change to the current retinal images compared to before these changes happened to the AR, VR or MR device. Retinal vasculature images can be used to measure the level of image distortion within an imaging system by resolving a specific pattern of high contrast lines. By processing the retinal images or their panoramic mosaic into binary (black and white) high contrast by intensity thresholding and/or segmentation, the vascular network can be made into a RetinaTest Target.
[0110] By measuring the change in the images of the RetinaTest Target before and after a change in performance of the AR, MR or MR device, a calibration measurement of imaging performance can be made dynamically. This calibration measurement can be transmitted to a local computing device or to a remote location for analysis and diagnosis of the change of performance of the AR, VR or MR device. Furthermore, the calibration measurement can be updated when corrective actions are implemented within the AR, VR or MR device which can be used in a feedback loop as an error signal for the purpose of regaining optimal performance of the AR, VR or MR device. Since the blood has a distinct optical absorption spectrum in the arteries and veins and scattering differences can be determined, the calibrated imaging performance should be performed across the spectral range of visible to near infrared wavelengths being used by the AR, VR or MR device.
[0111] Gaze Tracking
[0112] The acquisition of template images and registration onto a baseline image as described above can be further used to determine the gaze position the user. In particular, as the user's gaze changes position, the angle between the optical axis of the camera and the fovea or other structures at the back of the eye will change accordingly, and by measuring the shift in this angle the gaze position can be determined.
[0113] While the above discussion has been directed primarily to detecting changes in the retina and monitoring for change, progression, occurrence etc. of eye disease, more generally the present methods can be used to monitor for other conditions (e.g., diabetes, etc.) that are not retinal conditions per se, but that may be measured in the retina. Furthermore, our methods can be also used to monitor improvement in a condition of the retina, for example, monitor effectiveness of a treatment or therapy, in addition to detecting onset or worsening of disease.
[0114] Other applications are of course possible as would be apparent to one skilled in the art.
[0115] The manuscript portion of our priority US provisional application includes data regarding experiments we conducted using our template matching method, including validation on a set of simulated images from the STARE dataset, and in-vivo templated images captured from the D-eye smartphone device matched to full fundus images and mosaicked full images. The interested reader is directed to that portion of the provisional application for further details.
[0116] As used in the claims, the term “head-worn retinal imaging device” is intended to refer broadly to any device worn or supported by the head which includes a detector or camera and associated optical components designed for imaging the retina, including but not limited to glasses, and augmented, mixed or virtual reality headsets. As another example, devices which include scanned light (from laser or LED) display using a near-infrared (NIR) wavelength can also be a camera with the addition of a fast NIR detector, and such a device could be adapted as a head-worn retinal imaging device.
[0117] The particulars shown herein are by way of example and for purposes of illustrative discussion of the preferred embodiments of the present invention only and are presented in the cause of providing what is believed to be the most useful and readily understood description of the principles and conceptual aspects of various embodiments of the invention. In this regard, no attempt is made to show structural details of the invention in more detail than is necessary for the fundamental understanding of the invention, the description taken with the drawings and/or examples making apparent to those skilled in the art how the several forms of the invention may be embodied in practice.
[0118] As used herein and unless otherwise indicated, the terms “a” and “an” are taken to mean “one”, “at least one” or “one or more”. Unless otherwise required by context, singular terms used herein shall include pluralities and plural terms shall include the singular. Unless the context clearly requires otherwise, throughout the description and the claims, the words ‘comprise’, ‘comprising’, and the like are to be construed in an inclusive sense as opposed to an exclusive or exhaustive sense; that is to say, in the sense of “including, but not limited to”. Words using the singular or plural number also include the plural and singular number, respectively. Additionally, the words “herein,” “above,” and “below” and words of similar import, when used in this application, refer to this application as a whole and not to any particular portions of the application.