SYSTEM AND METHOD FOR RETINA TEMPLATE MATCHING IN TELEOPHTHALMOLOGY
20240315553 ยท 2024-09-26
Assignee
Inventors
- Eric J. Seibel (Seattle, WA)
- Chen GONG (Seattle, WA, US)
- Steven L. BRUNTON (Seattle, WA, US)
- Nils Benjamin ERICHSON (Seattle, WA, US)
- Laura TRUTOIU (Seattle, WA, US)
- Brian T. Schowengerdt (Seattle, WA, US)
Cpc classification
A61B2576/02
HUMAN NECESSITIES
A61B3/0025
HUMAN NECESSITIES
G06T3/4038
PHYSICS
A61B3/14
HUMAN NECESSITIES
G06V10/76
PHYSICS
G06T2207/10101
PHYSICS
A61B3/12
HUMAN NECESSITIES
International classification
A61B3/12
HUMAN NECESSITIES
A61B3/00
HUMAN NECESSITIES
A61B3/14
HUMAN NECESSITIES
Abstract
A retina image template matching method is based on the registration and comparison between the images captured with portable low-cost fundus cameras (e.g., a consumer grade camera typically incorporated into a smartphone or tablet computer) and a baseline image. The method solves the challenges posed by registering small and low-quality retinal template images captured with such cameras. Our method combines dimension reduction methods with a mutual information (MI) based image registration technique. In particular, principle components analysis (PCA) and optionally block PCA are used as a dimension reduction method to localize the template image coarsely to the baseline image, then the resulting displacement parameters are used to initialize the MI metric optimization for registration of the template image with the closest region of the baseline image.
Claims
1. A computer-implemented method for registering a narrow field of view template image to a wide field of view, previously obtained, baseline image, the method comprising: cropping the baseline image into a multitude of smaller offset target images; applying a dimension reduction method to map the offset target images to a representation in a lower dimensional space; mapping the template image into the lower dimensional space using the dimension reduction method; finding the corresponding nearest target image for the template image in the lower dimensional space; registering the template image to the nearest target image; identifying the location of the template image on the baseline image based on the position of the nearest target image; and registering the template image to the baseline image at the identified location.
2. The method of claim 1, wherein the baseline image comprises a fundus image.
3. The method of claim 2, wherein the template image comprises an image captured by a portable fundus camera.
4. The method of claim 3, wherein the portable fundus camera comprises a camera embodied in a smartphone or tablet computer configured with apparatus to assist in taking a photograph of the eye.
5. The method of claim 4, wherein the cropping, applying, mapping, finding, registering, identifying, and registering are performed in a processing unit in the smartphone or tablet computer.
6. The method of claim 2, wherein the fundus image is obtained without chemical dilation of the pupil of the subject.
7. The method of claim 1, wherein registering the template image to the nearest target image employs a mutual information procedure.
8. The method of claim 1, wherein applying a dimension reduction method to map the offset target images to a representation in a lower dimensional space and mapping the template image into the lower dimensional space using the dimension reduction method comprises Principal Component Analysis.
9. The method of claim 1, wherein finding the corresponding nearest target image for the template image in the lower dimensional space is performed using block Principal Component Analysis.
10. The method of claim 1, further comprising determining the gaze position of the subject.
11. The method of claim 1, further comprising locating a surgical tool in the eye from the registered template images.
12. An extended reality device, comprising: an imaging device; a processor operatively coupled to the imaging device; and a memory storing therein a sequence of instructions which, when executed by the processor, causes the processor to perform a set of acts for registering a narrow field of view template image to a wide field of view, previously obtained, baseline image, the set of acts comprising: cropping the baseline image into a multitude of smaller offset target images; applying a dimension reduction method to map the offset target images to a representation in a lower dimensional space; mapping the template image into the lower dimensional space using the dimension reduction method; finding the corresponding nearest target image for the template image in the lower dimensional space; registering the template image to the nearest target image; identifying the location of the template image on the baseline image based on the position of the nearest target image; and registering the template image to the baseline image at the identified location.
13. The device of claim 12, wherein the baseline image comprises a fundus image.
14. The device of claim 13, wherein the template image comprises an image captured by a portable fundus camera.
15. The device of claim 14, wherein the portable fundus camera comprises a camera embodied in a smartphone or tablet computer configured with apparatus to assist in taking a photograph of the eye.
16. The device of claim 15, wherein the cropping, applying, mapping, finding, registering, identifying, and registering are performed in a processing unit in the smartphone or tablet computer.
17. The device of claim 13, wherein the fundus image is obtained without chemical dilation of the pupil of the subject.
18. The device of claim 12, wherein registering the template image to the nearest target image employs a mutual information procedure.
19. The device of claim 12, wherein applying a dimension reduction device to map the offset target images to a representation in a lower dimensional space and mapping the template image into the lower dimensional space using the dimension reduction device comprises Principal Component Analysis.
20. A non-transitory machine accessible storage medium having stored thereupon a sequence of instructions which, when executed by a processor of a mixed reality device, causes the processor to perform a set of acts for registering a narrow field of view template image to a wide field of view, previously obtained, baseline image, the set of acts comprising: cropping the baseline image into a multitude of smaller offset target images; applying a dimension reduction method to map the offset target images to a representation in a lower dimensional space; mapping the template image into the lower dimensional space using the dimension reduction method; finding the corresponding nearest target image for the template image in the lower dimensional space; registering the template image to the nearest target image; identifying the location of the template image on the baseline image based on the position of the nearest target image; and registering the template image to the baseline image at the identified location.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0029] The appended drawing figures are offered by way of example and not limitation of currently preferred embodiments of this disclosure.
[0030]
[0031]
[0036]
[0037]
[0038]
[0039]
[0040]
DETAILED DESCRIPTION
[0041] This document discloses an efficient and accurate retinal matching system and method combining dimension reduction and mutual information (MI), we refer to the technique here as RetinaMatch. By way of overview, the dimension reduction initializes the MI optimization as a coarse localization process, which narrows the optimization domain and avoids local optima. The disclosed system and method outperforms to-date existing template matching solutions. In addition, we disclose a system and method for image mosaicking with area-based registration, providing a robust approach which may be used when the feature-based methods fail. To the best of our knowledge, this is the first template matching technique for retina images with small template images from unconstrained retinal areas.
[0042] Our approach is an improvement over the area-based matching methods with MI metric, since it is more reliable to achieve accurate and robust template matching near the alignment position. One unique aspect of our approach is that we combine dimension reduction methods with the MI-based registration to reduce the interference of local extrema and improve the matching efficiency.
[0043] An example of the practical use of our method in monitoring the retina in a teleophthalmology setting is shown in
[0044] It is also contemplated that the template images captured by the smartphone 12 could be sent over the network 16, 18 to a computing system 22 in the eye clinic and the processing steps of
[0045] Specific examples of the applications of the retinal template matching in a teleophthalmology setting will be discussed at length later in this document.
[0046] With the above description in mind, one of the principal aspects of this disclosure is application of dimension reduction methods with the MI-based registration to reduce the inference of local extrema and improve the matching efficiency.
[0047]
[0048] The process of
[0049] The procedures shown in panels (a) and (b) of
[0050] With the above explanation in mind, attention will now be directed to
[0051]
[0052] A specific embodiment of our procedure of
1. FIG. 2, Panel (a): Create Target Images from the Full (Baseline) Image
[0053] We define the full image and template as F and S respectively. The full image F is split into target images I.sub.1, I.sub.2, . . . , I.sub.N:
I.sub.i=?(b.sub.i, F).
[0054] The function ? crops the target images I.sub.i from F at b.sub.i and b.sub.i=[x.sub.i, y.sub.i, h, w], where (x.sub.i, y.sub.i) denotes the center position and (h,w) denotes the width and height of the cropped image. There is a certain displacement, f, of neighboring target images in the x and y axes. As shown in
[0055] Target images are resized to vectors and form the matrix X ?.sup.n?d.
2. FIG. 2 Panel (b): Create Low Dimensional Representations of the Target Images with PCA
[0056] Dimension reduction methods allow the construction of low-dimensional summaries, while eliminating redundancies and noise in the data. To estimate the template location in the 2d space, the full image dimension is redundancy, thus we apply dimension reduction methods for the template coarse localization.
[0057] Generally, we can distinguish between linear and nonlinear dimension reduction techniques. The most prominent linear technique PCA. PCA is selected as the dimension reduction method in RetinaMatch since PCA is simple and versatile. Specifically, PCA forms a set of new variables as a weighted linear combination of the input variables. Consider a matrix X=[X.sub.1, x.sub.2, . . . , X.sub.d] of dimension n?d, where n denotes the number of observations and d is the number of variables. Further, we assume that the matrix X is column-wise mean centered. The idea of PCA is to form a set of uncorrelated new variables (referred to as principal components) as a linear combination of the input variables:
where z.sub.i is the .sub.ith principal component (PC) and w.sub.i is the weight vector. The first PC explains most of the variation in the data, the subsequent PCs then account for the remaining variation in descending order. Thereby, PCA imposes the constraint that the weight vectors are orthogonal. This problem can be expressed compactly as the following minimization:
where ?.?.sub.F is the Frobenius norm. The weight matrix W that maps the input data to a subspace turns out to be the right singular vectors of the input matrix X. Often a low-rank approximation is desirable, e.g., we compute the k dominant PCs using a truncated weight matrix W.sub.k=[w.sub.1,w.sub.2, . . . , w.sub.k]. k is some integer, such as 20.
[0058] PCA is generally computed by the singular value decomposition (SVD). Many algorithms have been developed to streamline the computation of the SVD or PCA for high-dimensional data that exhibits low-dimensional patterns, see J. N. Kutz, et al., Dynamic mode decomposition: data-driven modeling of complex systems. SIAM, 2016, vol. 149. In particular, tremendous strides have been made accelerating the SVD and related computations using randomized methods for linear algebra. See the references 24-31 cited in the manuscript portion of the priority U.S. provisional application. Since we have demonstrated high performance with less than 20 principal components, the randomized SVD is used to compute the principal components, improving the efficiency in this retinal mapping application for mobile device platforms (e.g., smartphone, tablet). The randomized algorithm proceeds by forming a sketch Y of the input matrix
Y=X?,
Where ? is a d?l random test matrix, say with independent and identically distributed random standard normal entries. Thus, the l columns of Y are formed as a randomly weighted linear combination of the columns of the input matric, providing a basis for the column space of X. Note, that l is chosen to be slightly larger than the desired number of principal components. Next, we form an orthonormal basis Q using the QR decomposition Y=QR. Now, we use this basis matrix to project the input data matrix to low-dimensional space
B=Q.sup.TX.
[0059] This smaller matrix B of dimension l?d can then be used to efficiently compute the low-rank SVD and subsequently the dominant principal components. Given the SVD of B=U?V.sup.T, we obtain the approximate principal components as
Z=QU?=XV.
[0060] Here, U and V are the left and right singular vectors and the diagonal elements of ? are the corresponding singular values. The approximation accuracy can be controlled via additional oversampling and power iterations.
[0061] Referring again to panel (b) of
Z=XW,
Where Z=[z.sub.1, z.sub.2, z.sub.3, . . . , z.sub.N].sup.T?.sup.n?l, W?
.sup.d?l and l <<d. The image space ?.sub.1 is mapped to a low-dimensional space ?.sub.2 with the mapping W. W and Z are saved in memory, in what we have called a dictionary, D.
[0062] It is important to note that PCA is sensitive to outliers, occlusions, and corruption in the data. In ophthalmological imaging applications, there are several potential sources of corruption and outliers when imaging the full image, including blur, uncorrected astigmatism, inhomogeneous illumination, glare from crystalline lens opacity, internal reflections (e.g., from the vitreoretinal interface and lens), transient floaters in the vitreous, and shot noise in the camera. Further, there is often a trade-off between illumination and image quality, and there is strong motivation to introduce as little light as necessary for the patient comfort and health. The robust principal component analysis (RPCA) was introduced specifically to address this issue, decomposing a data matrix into the sum of a matrix containing low-rank coherent structure and a sparse matrix of outliers and corrupt entries. In general, RPCA is more expensive than PCA, requiring an iterative optimization to decompose the original matrix into sparse and low-rank components. Each step of the iteration is as expensive as regular PCA, and typically on the order of tens of iterations are required; however, PCA may be viewed as an offline step in our procedure, so that this additional computational cost is manageable. RPCA has been applied with success in retinal imaging applications to improve image quality. In the examples presented in this work, the data appears to have few enough outliers so that RPCA is not necessary, although it is important to keep RCPA as an option for data with outliers and corruption. Further details on RPCA are contained in the references cited in the manuscript portion of our prior provisional application.
3. FIG. 2 Panel (c): Coarse LocalizationFind the Nearest Target Image in the Low-Dimensional Space
[0063] Given a template S, the coarse position can be estimated by recognizing its nearest target image. The nearest target image in the image space ?.sub.1 should also be the nearest representation of S in the lower dimensional space ?.sub.2. Accordingly, we obtain the low- dimension feature z.sub.s of the template in ?.sub.2:
z.sub.s={tilde over (S)}W,
where s?.sup.d is the reshaped vector of template S. Let ?(z.sub.s, z) be the Euclidean distance between z.sub.s and a feature z in Z. z* is the nearest target feature of the source image S in ?.sub.2:
z*=arg.sub.zmin?(z.sub.s, z).
[0064] The corresponding target image location is used as the coarse location of S. Ideally, the difference between the coarse location and the ground truth in x and y axes should be less than ?/2 pixels.
[0065] In one of the experiments we performed, PCA outperforms other non-linear dimension reduction methods, while the error is larger than ?/2. The main reason is that the image degradation creates spurious features that contribute to the final classification. To reduce the influence of local features, we implement block PCA to further improve the accuracy of the coarse localization. By computing the PCA of different local patches in the template, the effect of local features, which cannot be located correctly, is reduced. This procedure is shown in
[0066] Obtaining the nearest target image, we crop a larger image at the same position from the full image as the new target image I. In this way, the template can have more overlap with the new target image when there is a large offset between two images. We segment I and the template S into small patches with the cropping function {tilde over (?)}, where the patch size is smaller than the source image with the axial displacement of neighboring patches ? Similarly, all image patches from I are mapped into the low-dimensional space ?.sub.3 with W. Let Z denote the low-dimensional representation of the target image distribution. Each template patch is then mapped to the space with W. The nearest target patch for each template patch is determined with the Euclidean distance as described before. We use the same weight for each region of the template for localization. Let b.sub.m be the mean of the coordinates of selected nearest target patches, which then represents the center of the template on I. Accordingly, the template location on the full image can be estimated and the region is cropped as the image ?. We store the representation of each of the target image patches in lower dimensional space in memory, referred to as dictionary T. The accurate registration is then applied to the template S and image ?. In this way the coarse localization provides an estimate of a good initial point for the accurate registration (panel (d) of
[0067] In the implementation of the proposed coarse localization, the full (baseline) image is assumed to exist so the dictionary D and dictionary T for each target image can be built in advance. This is the pre-computed part as shown in
[0068]
Example Processing Instructions for Coarse Localization:
[0069]
TABLE-US-00001 1 Map template S into space ?.sub.2: z.sub.s = {tilde over (s)}W. 2 Determine closest target image I with corresponding z*: z* = arg min.sub.z ?(z.sub.s, z). z* ? Z. 3 Segment S into [S.sub.p.sup.1, S.sub.p.sup.2, . . . , S.sub.p.sup.n]: S.sub.p.sup.i = {tilde over (?)}(b.sub.i, S): Segment I into [I.sub.p.sup.1, I.sub.p.sup.2, . . . , I.sub.p.sup.n]: I.sub.p.sup.i = {tilde over (?)}(b.sub.i, I): 4 Map target patches I.sub.p.sup.i into space ?.sub.3: Z = I.sub.pW, where I.sub.p is formed with vectorized I.sub.p.sup.i. 5 For each template patch S.sub.p.sup.i: 6 (i)Map S.sub.p.sup.i into space ?.sub.3: {tilde over (z)}.sub.s.sup.i = S.sub.p.sup.iW. 7 (ii)Determine its closest target patch I.sub.p.sup.Idx(i) with index Idx (i). 8
4. FIG. 2 Panel (d) Accurate Registration and Location of Template Onto Baseline
[0070] Panel (d) of
(1) Image Registration Between Template and Nearest Target Image Using Mutual Information (MI) (FIG. 4 step 310)
[0071] In this section, we describe the maximization of MI for multimodal image registration. We define images S and ? as the template and target images, respectively. A transform u is defined to map pixel locations x?S to pixel locations in ?.
[0072] The main idea of the registration is to find a deformation ? at each pixel location x that maximizes the MI between the deformed template image S(u(x)) and the target image ?(x). Accordingly,
Here, i.sub.1 and i.sub.2 are the image intensity values in S(u(x)) and ?(x), respectively and p(i.sub.1) and p(i.sub.2) are their marginal probability distributions while p(i.sub.1, i.sub.2) is their joint probability distribution. The probability distributions p(i.sub.1, i.sub.2) reflect the degree to which the greyscale (image intensity) values of each pixel in S(u(x)) and ?(x) are similar; p(i.sub.1, i.sub.2) has a high value (closer to 1) if the pixel values are similar, and low value (closer to 0) if the pixel values are dissimilar. In more detail, in terms of mutual information, based on discrete data like images, each pixel has a grayscale value for 0 to 255. (Although examples herein may describe use of grayscale images for the fundus image work, embodiments are not so limited and may also employ color images as appropriate). We first compute the joint histogram of two images: the joint histogram will be 256x256, which counts the number of corresponding pixels' grayscale from two images. For example, if in the first pixel, one image has a grayscale of 100 and another one is 120, then the joint histogram map (100, 120) will add one. After we finished the joint histogram, the joint probability p(i.sub.1, i.sub.2) can be obtained by normalizing the joint histogram. Then the marginal probability is computed according to:
(2) Locate the Template Onto the Full Image (FIG. 4 step 312)
[0073] In this step images S and ? are accurately registered with maximization of mutual information, as per sub-step (d)(1) above. The location of mage ? on the full image F becomes the estimated displacement of the template S. In our work, the transform u for alignment is given as an affine transformation:
[0074] It will be appreciated that the processing to create the target images and map them into lower dimensional space (panels (a) and (b) of
Image Mosaicking
[0075]
[0076] As pointed out previously, the full retina image can be stitched into a panorama by using many small templates. Users must capture a series of images in naturally unconstrained eye positions to explore different regions of the retina. It is problematic to determine adjacent images before the registration when we apply area-based registration approaches, because at that time they may not have effective descriptors for matching.
[0077] Related to the dimension reduction in the proposed template matching method, here we present the procedure shown in the table below to learn the positional relationship of images to be stitched. In this way, the adjacent images can be recognized and registered efficiently.
[0078] For a series of small images X.sub.i, we form the matrix X. PCA is applied to X and returns the low-dimensional features for each image in ?.sub.2. The distance between features in ?.sub.2 indicates the distance between images. We find the nearest N (e.g., N=3) target neighbors in the low dimensional space. The nearest neighbor X.sub.j of image X.sub.i is the one with largest overlap; the image pair is then registered with MI-based approach. To improve the algorithm robustness, the first N nearest neighbors for each image are first selected to compute MI with, and we keep the one with the largest metric value. The above procedure can be represented in the following pseudocode.
Processing Instructions: Image Stitching (with Reference to FIG. 6)
[0079] 1 Map images into space ?.sub.2:Z=XW. (step 502) [0080] 2 For each image X.sub.i: [0081] 3 (i).Find the nearest N (e.g., N=3) neighbors X.sub.j minimizing the feature distance ?(Z.sub.i,Z.sub.j). (step 504) [0082] 4 (ii).Compute the Mutual Information between each X.sub.j and X.sub.i and take the adjacent image with highest MI. (step 506, 508) [0083] 5 Panorama R Mosaicking: Align all the adjacent images with mutual information based registration method. (step 510) [0084] 6 Panorama blending. (step 512) [0085] 7 return mosaicked panorama R. (step 514)
APPLICATIONS
[0086] Our method of template matching with baseline images and image mosaicking allows for longitudinal comparisons with previously obtained fundus images of the patient. Such longitudinal comparisons have several applications in the field of ophthalmology as will be described below. Such applications are examples of how the methods of this disclosure can be practiced in a teleophthalmology setting. Other suitable applications are also supported by the embodiments described herein, including options outside of the field of retinal template matching.
Hypertension
[0087] In the retinal symptom of hypertension, the larger arteries constrict and the venous vessels enlarge in diameter. Ophthalmologists can select several detection points on the vessels. With the captured images coming from the patient as per
Abusive Head Trauma
[0088] The biomarkers of abusive head trauma (AHT) is another example. The most common retinal manifestation of AHT is multiple retinal hemorrhages in multiple layers of the retina. Matching the captured images onto the full retina image, the hemorrhagic spots can be easily segmented after the subtraction of the current retina regions and previous status. The AHT then can be recognized automatically when such spots are detected. This method permits identification of AHT from images obtained with portable fundus cameras.
Diabetic Retinopathy
[0089] The obvious symptoms of diabetic retinopathy (DR) are retina hemorrhages and the presence of exudate. They can be monitored follow the similar process of AHT screening.
Glaucoma
[0090] Glaucoma can cause the optic nerve cup to enlarge. Our matching method can automatically select the images that cover the optic nerve. The following segmentation can be easily implemented and a computation of the optic cup diameter performed. Enlargement of the optic nerve cup over time can be ascertained by comparing the computations from a current image with an image from a previous point in time.
Use RetinaMatch as a General Image Template Matching Method
[0091] Besides the retina images, the technique of RetinaMatch can be used in other type of image template matching tasks. Note that our method of
Use RetinaMatch for Camera Localization
[0092] Having the image of the full view, our method of
Augmented Reality (AR), Eye Glasses, etc. and Monitoring Changes Over Time
[0093] A retinal imaging system (e.g., consumer grade camera with ancillary imaging device, e.g., D-eye) can be portable and further, can be worn as integrated into, for example, glasses, or an Augmented Reality (AR), Virtual Reality (VR) and/or Mixed Reality (MR) headset, allowing a series of images to be taken and analyzed, either daily, weekly, monthly, or when the user or ophthalmologist requests. These measurements can be discrete, continual, but in a time series and analyzed longitudinally over the increasing time period. Change in a retina can detected by registering and comparing the captured small FOV images to a full baseline retina image using our template matching method.
[0094] AR, VR and/or MR devices can be used to optically scan the retina to form images and thereby acquire the template images. Even more pragmatically, spectacles or sunglasses can be used because of the smaller size, lower costs, and increasing utility to the user. A scanned light beam entering the pupil of the eye and striking the retina to form video rate images perceived by the user's brain can also be used to acquire images of high contrast structures, such as the vasculature containing blood.
[0095] A device can operate without major changes in performance during its lifetime and can be used as a monitor of the condition of a user's eye. By comparing retinal images from such a device over time, the changes in the user's optical system (such as cornea, intraocular lens, retina, and liquid environments) can be monitored to alert the user in possible health changes. For example, these changes can be gradual, like increasing light scattering from the crystalline or intraocular lens due to cataract formation, or the appearance and structural changes in the retina due to diabetic retinopathy. In addition, chronic diseases which may have variations over time in the blood vessel size and shape in conditions of hypertension are another example. Acute changes such as bleeding within the retinal can indicate brain trauma. Relative and repeatable changes in number, size, and shape of structures in the retinal images may indicate that the measured change is due to a particular disease type and not that the AR, VR, MR, glasses, or other type of monitoring device has slowly or suddenly changed its imaging performance or has become unstable.
[0096] However, in many healthy users the optical system will be unchanging over time. In this case, the vasculature of the retina can be used as a test target for detecting optical misalignments, focus errors, light scanning errors and distortions, non-uniform and color imbalance in the illumination, and aberrations in the imaging system. This situation can occur if the monitoring device, such as an AR, VR, or MR device is degraded due to mechanical impact, breakage, applied stresses, applied vibration, thermal changes, and opto-electrical disruption or interference. These changes can be observed in a measurable change to the current retinal images compared to before these changes happened to the AR, VR or MR device. Retinal vasculature images can be used to measure the level of image distortion within an imaging system by resolving a specific pattern of high contrast lines. By processing the retinal images or their panoramic mosaic into binary (black and white) high contrast by intensity thresholding and/or segmentation, the vascular network can be made into a RetinaTest Target.
[0097] By measuring the change in the images of the RetinaTest Target before and after a change in performance of the AR, MR or MR device, a calibration measurement of imaging performance can be made dynamically. This calibration measurement can be transmitted to a local computing device or to a remote location for analysis and diagnosis of the change of performance of the AR, VR or MR device. Furthermore, the calibration measurement can be updated when corrective actions are implemented within the AR, VR or MR device which can be used in a feedback loop as an error signal for the purpose of regaining optimal performance of the AR, VR or MR device. Since the blood has a distinct optical absorption spectrum in the arteries and veins and scattering differences can be determined, the calibrated imaging performance should be performed across the spectral range of visible to near infrared wavelengths being used by the AR, VR or MR device.
Gaze Tracking
[0098] The acquisition of template images and registration onto a baseline image as described above can be further used to determine the gaze position the user. In particular, as the user's gaze changes position, the angle between the optical axis of the camera and the fovea or other structures at the back of the eye will change accordingly, and by measuring the shift in this angle the gaze position can be determined.
[0099] While the above discussion has been directed primarily to detecting changes in the retina and monitoring for change, progression, occurrence etc. of eye disease, more generally the present methods can be used to monitor for other conditions (e.g., diabetes, etc.) that are not retinal conditions per se, but that may be measured in the retina. Furthermore, our methods can be also used to monitor improvement in a condition of the retina, for example, monitor effectiveness of a treatment or therapy, in addition to detecting onset or worsening of disease.
[0100] Other applications are of course possible as would be apparent to one skilled in the art.
[0101] The manuscript portion of our priority U.S. provisional application includes data regarding experiments we conducted using our template matching method, including validation on a set of simulated images from the STARE dataset, and in-vivo templated images captured from the D-eye smartphone device matched to full fundus images and mosaicked full images. The interested reader is directed to that portion of the provisional application for further details.
[0102] As used in the claims, the term head-worn retinal imaging device is intended to refer broadly to any device worn or supported by the head which includes a detector or camera and associated optical components designed for imaging the retina, including but not limited to glasses, and augmented, mixed or virtual reality headsets. As another example, devices which include scanned light (from laser or LED) display using a near-infrared (NIR) wavelength can also be a camera with the addition of a fast NIR detector, and such a device could be adapted as a head-worn retinal imaging device.
[0103] The particulars shown herein are by way of example and for purposes of illustrative discussion of the preferred embodiments of the present invention only and are presented in the cause of providing what is believed to be the most useful and readily understood description of the principles and conceptual aspects of various embodiments of the invention. In this regard, no attempt is made to show structural details of the invention in more detail than is necessary for the fundamental understanding of the invention, the description taken with the drawings and/or examples making apparent to those skilled in the art how the several forms of the invention may be embodied in practice.
[0104] As used herein and unless otherwise indicated, the terms a and an are taken to mean one, at least one or one or more. Unless otherwise required by context, singular terms used herein shall include pluralities and plural terms shall include the singular. Unless the context clearly requires otherwise, throughout the description and the claims, the words comprise, comprising, and the like are to be construed in an inclusive sense as opposed to an exclusive or exhaustive sense; that is to say, in the sense of including, but not limited to. Words using the singular or plural number also include the plural and singular number, respectively. Additionally, the words herein, above, and below and words of similar import, when used in this application, refer to this application as a whole and not to any particular portions of the application.