Method of Diagnosis
20230047141 · 2023-02-16
Inventors
Cpc classification
G06F18/2414
PHYSICS
G16H50/20
PHYSICS
G06V20/69
PHYSICS
International classification
G06V20/69
PHYSICS
Abstract
The invention relates to methods for determining the stage of a disease, particularly an ocular neurodegenerative disease such as Alzheimer's, Parkinson's, Huntington's and glaucoma, comprising the steps of identifying the status of microglial cells in the retina and relating that status to disease stage. Methods for identifying cells in the eye are also provided, as are labelled markers and the use thereof.
Claims
1. A method of determining the stage of a disease, especially a neurodegenerative disease, comprising the steps of: a) generating an image of the activation status of microglia cells in a subject's eye and b) relating the status of the cells to disease stage.
2. The method of claim 1, further comprising one or both of steps: c) counting the number of activated, ramified and/or amoeboid microglia in the image generated; and d) comparing the number or percentage of activated, ramified or amoeboid microglia cells found in the image with a previously obtained image, or with the expected the number or percentage of activated, ramified or amoeboid microglia.
3. The method according to claim 1 or claim 2, further comprising the step of identifying a pattern of cell status in the eye and relating that pattern to disease state.
4. The method according to any preceding claim, wherein the subject is a subject to whom a labelled marker has been administrated.
5. The method according to any of claims 1 to 3, further comprising the step of administering a labelled marker to the subject.
6. The method of claim 4 or 5, wherein the labelled marker is an apoptotic marker, particularly a labelled annexin, more particularly annexin 5.
7. The method of claim 4, 5 or 6, wherein the label is a visible label, particularly a. wavelength-optimised label, more particularly D-776.
8. The method of any preceding claims, further comprising the step of generating an image of apoptosing cells; and, optionally, counting the number of apoptosing cells and/or observing the pattern of apoptosing cells; and, optionally, comparing the number or pattern of apoptosing cells with the expected number or pattern or with the number or pattern of apoptosing cells in an image previously generated from the subject.
9. The method of any preceding claim, further comprising one or more of the following steps: comparing the image with an image or with more than one image of the subject's eye obtained at an earlier time point; comparing the number or pattern of activated and/or amoeboid microglia in one image with a previous image; comparing specific cells in one image with the same cells in a previous image; and comparing the number or pattern of apoptosing cells or comparing specific cells in one image with the same cells in an earlier image.
10. The method of claim 9, comprising the step of overlaying one image with one, two, three or more additional images.
11. The method of any preceding claim, wherein the disease is an ocular neurodegenerative disease.
12. The method of any preceding claim, further comprising the step of determining an appropriate treatment for the subject and/or administering to the subject a treatment, particularly for glaucoma or another neurodegenerative disease.
13. A labelled apoptotic marker for use in identifying microglia activation status.
14. A method of identifying cells in an image of the retina, comprising the steps of: a) providing an image of a subject's retina; b) identifying one or more spots on each image as a candidate of a labelled cell; c) filtering selections; and, optionally, d) normalising the results for variations in intensity.
15. The method of claim 14, further comprising the step of providing more than one image of the subject's retina.
16. The method of claim 15, further comprising the step of aligning the images to ensure cells seen in one image are aligned with cells seen in the other image.
17. The method of claim 16, further comprising the step of accounting for known variants that may cause false candidate identification.
18. The method of any of claims 14 to 17, wherein at least one of the steps is carried out by an automated means, and, optionally, wherein the automated means is trained to improve results going forward.
19. The method of any of claims 14 to 18, wherein the labelled cells are microglia cells; retinal nerve cells, especially retinal ganglion cells, or both.
20. A computer-implemented method of identifying the status of cells in the retina to, for example, determine the stage of a disease, the method comprising: a) providing an image of a subject's retina; b) identifying one or more spots on each image as a candidate of a labelled cell; c) filtering selections made by template matching using an object classification filter; and, optionally, d) normalising the results for variations in intensity.
21. A computer program for identifying the status of cells in the retina to, for example, determine the stage of a disease which, when executed by a processing system, causes the processing system to: a) provide an image of a subject's retina; b) use template mapping to identify one or more spots on each image as a candidate of a labelled cell; c) filter selections made by template matching using an object classification filter; and d) normalise the results for variations in intensity.
22. A non-transitory computer-readable storage medium comprising a set of computer-readable instructions stored thereon, which, when executed by a processing system, cause the processing system to perform the method of any one of claims 14 to 19.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0037]
[0038]
[0039]
[0040]
[0041]
[0042]
[0043]
[0044]
[0045]
[0046] CNN training (A) and validation (B) curves. A good accuracy is achieved in 200 epochs (training cycles) although training was left for 300 epochs to verify stability. The matching validation accuracy also shows similar accuracy without signs of over training. The accuracy was found to be 97%, with 91.1% sensitivity and 97.1% specificity.
[0047]
[0048]
[0049]
DESCRIPTION OF THE INVENTION
Example 1
[0050] Labelled annexin V was prepared as described in WO2009077750A1. The labelled annexin was administered as described in Cordeiro M F, Guo L, Luong V, et al. Real-time imaging of single nerve cell apoptosis in retinal neurodegeneration. Proc Natl Acad Sci USA 2004; 101: 13352-13356.
[0051] Iba-1 (Ionized calcium binding adaptor molecule 1 (Iba1) was used as a marker for microglia, using techniques known in the art.
[0052] Brn3a was used as a marker for retinal ganglion cells, using techniques known in the art.
[0053] Animal used included naïve rats, glaucoma model rats (OHT), Alzheimer's model mice, and glaucoma model mice. Such models are known in the art. Examples are described in WO2011055121A1.
[0054]
[0055] In
[0056] As can be seen in
[0057] As can be seen from
Example 2
[0058] Artificial intelligence is increasingly used in healthcare, especially ophthalmology. (Poplin et al., 2018)(Ting et al., 2019) Machine learning algorithms have become important analytical aids in retinal imaging, being frequently advocated in the management of diabetic retinopathy, age-related macular degeneration and glaucoma, where their utilization is believed to optimise both sensitivity and specificity in diagnosis and monitoring. (Sebastian A Banegas et al., 2015)(Quellec et al., 2017; Schmidt-Erfurth, Bogunovic, et al., 2018; Schmidt-Erfurth, Waldstein, et al., 2018; Orlando et al., 2019) The use of deep learning in these blinding conditions has been heralded as an advance to reduce their health and socio-economic impact, although their accuracy is confounded by dataset size and deficient reference standards. (Orlando et al., 2019)
[0059] Glaucoma is a progressive and slowly evolving ocular neurodegenerative disease that it is the leading cause of global irreversible blindness, affecting over 60.5 million people, predicted to double by 2040, as the aging population increases. (Quigley and Broman, 2006; Tham et al., 2014) A key objective in glaucoma research over the last few years is to identify those at risk of rapid progression and blindness. This has included, methods involving multiple levels of data including structural (optical coherence tomography (OCT), disc imaging) and functional (visual fields or standard automated perimetry (SAP)) assessments. However, several studies have demonstrated there is great variability amongst clinicians in agreement over progression using standard assessments including SAP, OCT and optic disc stereo photography. (A C Viswanathan et al., 2003)(Moreno-Montañés et al., 2017)(Sebastian A Banegas et al., 2015) However, clinical grading is regarded as the gold standard in real world practice and in deep learning datasets. (Jiang et al., 2018; Kucur, Hollo and Sznitman, 2018; Asaoka et al., 2019a; Ian J C MacCormick et al., 2019; Medeiros, Jammal and Thompson, 2019a; Thompson, Jammal and Medeiros, 2019; Wang et al., 2019) Moreover, it is recognised that both OCT
and SAP change only after significant death of a large number of retinal ganglion cells (RGC), (Harwerth et al., 2007) and with this the unmet need for earlier markers of disease.
[0060] Recently, we reported a novel method to visualise apoptotic retinal cells in the retina in humans called DARC (Detection of Apoptosing Retinal Cells). (Cordeiro et al., 2017) The molecular marker used in the technology is fluorescently labelled annexin A5, which has a high affinity for phosphatidylserine exposed on the surface of cells undergoing stress and in the early stages of apoptosis. The published Phase 1 results suggested that the number of DARC positively stained cells seen in a retinal fluorescent image could be used to assess glaucoma disease activity, but also correlated with future glaucoma disease progression, albeit in small patient numbers. DARC has recently been tested in more subjects in a Phase 2 clinical trial (ISRCTN10751859).
[0061] Here we describe an automatic method of DARC spot detection which was developed using a CNN, trained on a control cohort of subjects and then tested on glaucoma patients in the Phase 2 clinical trial of DARC. CNNs have shown strong performance in computer vision tasks in medicine, including medical image classification.
Materials and Methods
Participants
[0062] The Phase 2 clinical trial of DARC was conducted at The Western Eye Hospital, Imperial College Healthcare NHS Trust, as a single-centre, open-label study with subjects each receiving a single intravenous injection of fluorescent annexin 5 (ANX776, 0.4 mg) between 15 Feb. 2017 and 30 Jun. 2017. Both healthy and progressing glaucoma subjects were recruited to the trial, with informed consent being obtained according to the Declaration of Helsinki after the study was approved by the Brent Research Ethics Committee. (ISRCTN10751859).
[0063] All glaucoma subjects were already under the care of the glaucoma department at the Western Eye Hospital. Patients were considered for inclusion in the study if no ocular or systemic disease other than glaucoma was present and they had a minimum of three recent, sequential assessments with retinal optical coherence tomography (Spectralis SD OCT, software version 6.0.0.2; Heidelberg Engineering, Inc., Heidelberg, Germany) and standard automated perimetry (SAP, HFA 640i, Humphrey Field Analyzer; Carl Zeiss Meditec, Dublin, Calif.) using the Swedish interactive threshold algorithm standard 24-2. Patient eligibility was deemed possible if evidence of progressive disease in at least one eye of any parameter summarised in Table 1 & 2, was found to be present, where progression was defined by a significant (*p<0.05; **p<0.01) negative slope in the rate of progression (RoP). SAP parameters included the visual field index (VFI) and mean deviation (MD). OCT parameters included retinal nerve fibre layer (RNFL) measurements at three different diameters from the optic disc (3.5, 4.1, and 4.7 mm) and Bruch's membrane opening minimum rim width (MRW). Where it was not possible to use machine in-built software to define the rate of progression, due to the duration of the pre-intervention period of assessment, linear rates of change of each parameter with time were computed using ordinary least squares. (Wang et al., no date; Pathak, Demirel and Gardiner, 2013)
[0064] Healthy volunteers were initially recruited from people escorting patients to clinics and referrals from local optician services who acted as PICs. Healthy volunteers were also recruited from the Imperial College Healthcare NHS Trust healthy volunteers database. Potential participants were approached and given an invitation letter to participate. Participants at PICs who agreed to be contacted were approached by the research team and booked an appointment to discuss the trial. Enrolment was performed once sequential participants were considered eligible, according to the inclusion and exclusion criteria selected by the inventors. Briefly, healthy subjects were included if: there
was no ocular or systemic disease, as confirmed by their GP; there was no evidence of any glaucomatous process either with optic disc, RNFL (retinal nerve fibre layer) or visual field abnormalities and with normal IOP (intraocular pressure); and they had repeatable and reliable imaging and visual fields.
DARC Images
[0065] All participants received a single dose of 0.4 mg of ANX776 via intravenous injection following pupillary dilatation (1% tropicamide and 2.5% phenylephrine), and were assessed using a similar protocol to Phase 1. (Cordeiro et al., 2017) Briefly, retinal images were acquired using a cSLO (HRA+OCT Spectralis, Heidelberg Engineering GmbH, Heidelberg, Germany) with ICGA infrared fluorescence settings (diode laser 786 nm excitation; photodetector with 800-nm barrier filter) in the high resolution mode. Baseline infrared autofluorescent images were acquired prior to ANX776 administration, and then during and after ANX776 injection at 15, 120 and 240 minutes. Averaged images from sequences of 100 frames were recorded at each time point. All images were anonymised before any analysis was performed. For the development of the CNN-algorithm, only baseline and 120 minute images from control and glaucoma subjects were used.
[0066] The breakdown of the images analysed are shown in the “Consort” diagram in
Manual Observer Analysis
[0067] Anonymised images were randomly displayed on the same computer and under the same lighting conditions, and manual image review was performed by five blinded operators using ImageJ® (National Institutes of Mental Health, USA). (‘ImageJ’, no date) The ImageJ ‘multi-point’ tool was used to identify each structure in the image which observers wished to label as an ANX776 positive spot. Each positive spot was identified by a vector co-ordinate. Manual observer spots for each image were compared: spots from different observers were deemed to be the same spot if they were within 30 pixels of one another. Where there was concordance of two or more observers, this was used within the automated application as the criteria for spots used to train and compare the system.
Automated Image Analysis Overview (FIG. 7)
[0068] To detect the DARC labelled cells, candidate spots were identified in the retinal images, then classified as “DARC” or “not DARC” using an algorithm trained using the candidates and the spots identified by manual observers.
A) Image Optimisation
[0069] Images at 120 minutes were aligned to the baseline image for each eye using an affine transformation followed by a non-rigid transformation. Images were then cropped to remove alignment artefacts. The cropped images then had their intensity standardised by Z-Scoring each image to allow for lighting differences. Finally, the high-frequency noise was removed from the images with a Gaussian blur with a sigma of 5 pixels.
B) Spot Candidate Detection
[0070] Template matching, specifically Zero Normalised Cross-Correlation (ZNCC) is a simple method to find candidate spots. 30×30 pixel images of the spots identified by manual observers were combined using a mean image function to create a spot template. This template was applied to the retinal image producing a correlation map. Local maxima were then selected and filtered with thresholds for the correlation coefficient and intensity standard deviation (corresponding to the brightness of the spot). These thresholds were set low enough to include all spots seen by manual observers. Some of the manual observations were very subtle (arguably not spots at all) and correlation low for quite distinct spots due to their proximity to blood vessels. This means the thresholds needed to be set very low and produce many more spot candidates than manually observed spots (approximately 50-1).
[0071] As can be seen from
C) Spot Classification
[0072] To determine which of the spot candidates were DARC cells, the spots were classified using an established Convolutional Neural Network (CNN) called MobileNet v2. (Sandler et al., 2018; Chen et al., 2019; Pan, Agarwal and Merck, 2019; Pang et al., 2019) This CNN enables over 400 spot images to be processed in a single batch. This allows it to cope with the 50-1 unbalanced data since each batch should have about 4 DARC spots.
[0073] Although the MobileNet v2 architecture was used, the first and last layers were adapted. The first layer became a 64×64×1 input layer to take the 64×64 pixel spot candidate images (this size was chosen to include more of the area around the spot to give the network some context). The last layer was replaced with a dense layer with sigmoid activation to enable a binary classification (DARC spot or not) rather than multiple classification. An alpha value for MobileNet of 0.85 was found to work best, appropriately adjusting the number of filters in each layer
D) Training
[0074] Training was performed only on control eyes. Briefly, retinal images were randomly selected from 120 minute images of 50% of the control patients. The CNN was trained using candidate spots, marked as DARC if 2 or more manual observers observed the spot. 58,730 spot candidates were taken from these images (including 1022 2-agree manually observed DARC spots). 70% of these spots were used to train, and 30% to validate. The retinal images of the remaining 50% of control patients were used to test the classification accuracy (48610 candidate spots of which 898 were 2-agree manually observed).
[0075] The data was augmented to increase the tolerance of the network by rotating, reflecting and varying the intensity of the spot images. The DARC spots class weights were set to 50 for spots and 1 for other objects to compensate for the 50-1 unbalanced data.
[0076] The training validation accuracy converges, and the matching validation accuracy also shows similar accuracy without signs of over training. As the training curves show (see
[0077] Three training runs were performed, creating three CNN models. For inference, the three models were combined: each spot was classified based on the mean probability given by each of the three models.
E) Testing on Glaucoma DARC Images
[0078] Once the CNN-aided algorithm was developed, it was tested on the glaucoma cohort of patients in images captured at baseline and 120 minutes. Spots were identified by manual observers and the algorithm. The DARC count was defined as the number of a ANX776-positive spots seen in the retinal image at 120 minutes after baseline spot subtraction.
Glaucoma Progression Assessment
[0079] Rates of progression were computed from serial OCTs on glaucoma patient 18 months after DARC. Those patients with a significant (p<0.05) negative slope were defined as progressing compared to those without who were defined as stable. Additionally, assessment was performed by 5 masked clinicians using visual field, OCT and optic disc measurements.
Results
Patient Demographics
[0080] 60 glaucoma patients were screened according to set inclusion/exclusion criteria, from which 20 patients with progressing (defined by a significant (p<0.05) negative slope in any parameter in at least one eye) glaucoma underwent intravenous DARC. Baseline characteristics of these glaucoma patients are presented in Table 2. 38 eyes were eligible for inclusion, of which 3 did not have images available for manual observer counts, 2 had images captured in low resolution mode and another 2 had intense intrinsic autofluorescence. All patients apart from 2 were followed up in the Eye clinic, with data being available to perform a post hoc assessment of progression.
Testing of Spot Classification
[0081] The results in
[0082] The sensitivity and specificity were encouragingly high, especially as the manual observation data that it was trained and tested on had been shown to have high levels of inter-observer variation. Typical examples of images and manual observer/algorithm spots are shown in
Classification Testing in Glaucoma Cohort
[0083] Using only the OCT global RNFL rates of progression (RoP 3.5 ring) performed at 18 months to define progression, the glaucoma cohort was divided into progressing and stable groups. Clinical agreement was poor between observers, hence, the use of objective, simple and single OCT parameter. Those patients with a significant (p<0.05) negative slope were defined as progressing compared to those without who were defined as stable, and are detailed in Table 3a. Of the 29 glaucoma eyes analysed, 8 were found to be progressing and 21 stable, by this definition.
[0084] Using this definition of glaucoma progression, a Receiver Operating Characteristic (ROC) curve was constructed for both CNN-aided algorithm and manual observer 2-agree and shown in
DARC Counts as a Predictor of Glaucoma Progression
[0085] DARC counts in both stable and progressing glaucoma groups with the CNN-aided algorithm are shown in
Discussion
[0086] The main goal of glaucoma management is to prevent vision loss. As the disease progresses slowly over many years, current gold standards of assessing changes not only take a long time to develop, but also after significant structural and functional damage has already occurred (Cordeiro et al., 2017). There is an unmet need in glaucoma for reliable measures to assess risk of future progression and effectiveness of treatments (Weinreb and Kaufman, 2009, 2011). Here, we describe a new CNN-aided algorithm which when combined with DARC—a marker of retinal cell apoptosis, is able to predict glaucoma progression defined by RNFL thinning on OCT, 18 months later. This method when used with DARC was able to provide an automated and objective biomarker.
[0087] The development of surrogate markers has been predominantly is cancer where they are used as predictors of clinical outcome. In glaucoma, the most common clinical outcome measure is vision loss followed by a decrease in quality of life for assessing treatment efficacy. Surrogates should enable earlier diagnoses, earlier treatment, and also shorter, and therefore more economical clinical trials. However, to be a valid surrogate marker, the measures have to be shown to be accurate. For example, OCT, which is in widespread use has been found to have a sensitivity and specificity of 83% and 88% respectively for detecting significant RNFL abnormalities (Chang et al., 2009) in addition to good repeatability (DeLeon Ortega et al., 2007) (Tan et al., 2012). In comparison, our CNN algorithm had a sensitivity of 85.7% and specificity of 91.7% to glaucoma progression.
[0088] Although the Phase 1 results suggested there was some level of DARC being predictive, this was done on a very small dataset (Cordeiro et al., 2017) with different doses of Anx776 of 0.1, 0.2, 0.4 and 0.5 mg, with a maximum of 4 glaucoma eyes per group, of which there were only 3 in the 0.4 mg group. In this present study, all subjects received 0.4 mg Anx776, and 27 eyes were analysed.
[0089] In clinical practice, glaucoma patients are assessed for risk of progression based on establishing the presence of risk factors including: older age, a raised intraocular pressure (IOP, too high for that individual), ethnicity, a positive family history for glaucoma, stage of disease, and high myopia (Jonas et al., 2017). More advanced disease risks included a vertical cup:disc ratio >0.7, pattern standard deviation of visual field per 0.2 dB increase, bilateral involvement and disc asymmetry, as also the presence of disc haemorrhages and pseudexfoliation (Gordon et al., 2002, 2003; Budenz et al., 2006; Levine et al., 2006; Miglior et al., 2007). However, none of these can be used to definitely predict individual progression.
[0090] Objective assessment is increasingly recognised as being important in glaucoma, as there is variable agreement between clinicians, even with technological aids. Poor agreement has been shown with respect to defining progression in patients using visual fields, OCT and stereophotography (A. C. Viswanathan et al., 2003)(Sebastián A. Banegas et al., 2015)(Blumberg et al., 2016)(Moreno-Montañés et al., 2017). Indeed, for this study, we asked five masked senior glaucoma specialists (co-authors) to grade for progression of patients using their clinical judgement based on optic disc assessment, OCT and visual fields; unfortunately, there was variable agreement between them (unpublished data). For this reason, a single, objective metric (Tatham and Medeiros, 2017) of rate of progression was used to define the groups used to test the CNN-aided algorithm.
[0091] The analysis of progression was post-hoc, and there was no protocol guiding treating clinicians during the 18 month period of follow-up. Similar to the oral memantine trial, (Weinreb et al., 2018) management of patients, especially with regard to IOP lowering, was left to the discretion of glaucoma specialist, and following normal standard of care. However, despite this and using the OCT global RFNL 3.5 ring RoP, 8 of 29 eyes were progressing at 18 months.
[0092] The poor agreement between clinicians identifying progression has generated great interest in the last few years in the use of artificial intelligence to help aid glaucoma diagnosis and prognosis using AI with optic disc photographs (Jiang et al., 2018) (Ian J. C. MacCormick et al., 2019) (Thompson, Jammal and Medeiros, 2019), visual fields (Pang et al., 2019) (Kucur, Hollo and Sznitman, 2018) and OCT (Asaoka et al., 2019b) (Medeiros, Jammal and Thompson, 2019b). A recent study by Medeiros et al described an algorithm to assess fundus photographs based on predictions of estimated RNFL thickness, achieved by training a CNN using OCT RNFL thickness measurements (Medeiros, Jammal and Thompson, 2019b). At specificity of 95%, the predicted measurements had a sensitivity of 76% whereas actual SD OCT measurements had sensitivity of 73%. For specificity at 80%, the predicted measurements had sensitivity of 90% compared to OCT measurements which had sensitivity of 90%. The authors suggest their method could potentially could be used to extract progression information from optic disc photographs, but like our study, comment that further validation on longitudinal datasets is needed.
[0093] Template matching is routinely used for tracking cells in microscopy with similar assessment needed to analyse single cells in vivo longitudinally in this study. For template matching here, a 30×30 pixel template was used, for the CNN a 64×64 pixel image was used. The reason for this size difference is template matching is sensitive to blood vessels and so a small template is beneficial to reduce the likelihood of a blood vessel being included. For the CNN a larger image is useful to give the CNN more context of the area around the spot which may be useful in classification.
[0094] Although the algorithm performs well, providing a viable method to detect progressive glaucoma 18 months ahead of alternative methods, we believe there are areas where it can be optimised, some of which are described below.
[0095] Alternative classification algorithms to MobileNetV2 such as Support Vector Machines (SVMs) or Random Forests require “hand-crafted” features which are difficult to produce as they need to account for complexities caused by the image capture such as non-linear intensity variation, optical blur, registration blur and low light noise, as well as biological complexities such as the patterning in the choroidal vasculature, blood vessels, blur due to cataracts etc. The network has some biases to do with the intensity of the original retinal image. We believe we can improve results by looking at the intensity standardisation and augmenting the data by varying the intensity in ways more realistic with a larger dataset. The performance of other networks such as VGG16 were evaluated, at the time of writing MobiNetV2 was found to perform best. We are continuing to evaluate if this network is optimum for this need. In comparison, VGG16, an alternate CNN, would be limited to 64 spots in a batch which could mean a batch has no DARC spots in it which hinders training. We have an alternative method that detects and classifies spots in a single step using the detection and segmentation algorithm, YOLO3. We believe this may be a more efficient and effective method with more data, however at this stage the highest accuracy we have achieved with YOLO is not as good as the method outlined in this document.
Conclusion
[0096] This study describes a CNN-aided algorithm to analyse DARC as a marker of retinal cell apoptosis in retinal images in glaucoma patients. The algorithm enabled a DARC count to be computed which when tested in patients was found to successfully predict OCT RNFL glaucoma progression 18 months later. This data supports use of this method to provide an automated and objective biomarker with potentially widespread clinical applications.
TABLE-US-00001 TABLE 1 Glaucoma Eligibility (Exclusion/Inclusion Criteria Glaucoma) Subject ID Eligible eye Diagnosis 6 Both Primary Open Angle Glaucoma 7 Both Glaucoma suspect 9 Both Glaucoma suspect 11 Both Glaucoma suspect 13 Both Glaucoma suspect 17 Both Glaucoma suspect 18 Both Glaucoma suspect 21 Both Primary Open Angle Glaucoma 23 Both Primary Open Angle Glaucoma 25 Both Glaucoma suspect 31 Left Primary Open Angle Glaucoma 32 Both Primary Open Angle Glaucoma 38 Both Primary Open Angle Glaucoma 39 Both Glaucoma suspect 44 Both Primary Open Angle Glaucoma 45 Both Glaucoma suspect 52 Both Glaucoma suspect 61 Both Primary Open Angle Glaucoma 72 Left Glaucoma suspect 74 Both Glaucoma suspect
TABLE-US-00002 TABLE 1b Glaucoma characteristics on study entry Diagnosis n (%) Glaucoma 8 (40) Glaucoma suspect 12 (60) Ocular hypertension 0 (0) Total 20
TABLE-US-00003 TABLE 2 Baseline and Qualification Progression parameters Glaucoma Patients OCT RNFL RNFL RNFL SAP 3.5 4.1 4.7 MRW MD VFI Subject Eye μm/year μm/year μm/year μm/year dB/year %/year 6 R + + L 7 R + + L 9 R + L + 11 R + + L 13 R + + + L + 17 R + L + 18 R + + + L 21 R + + L + 23 R + + + + L 25 R + + L 31 R + L 32 R + + L 38 R + + + + L + + + 39 R + + L 44 R + L + 45 R + L 52 R + + + L + 61 R + + + + + + L + 72 R + + + L + 74 R + + + + L + +
TABLE-US-00004 TABLE 3a Progression classification per eye (OCT global RNFL 3.5 ring) 18 months after DARC Category Number of eyes Progressing 8 Stable 21 Unknown 4 N/A 5 Total 38
TABLE-US-00005 TABLE 3b Clinical findings of affected eyes meeting the inclusion criteria Glaucoma Healthy volunteer mean (SD) mean (SD) BCVA, logmar 0.01 (0.08) −0.03 (0.08) IOP, mmHg 18.90 (2.61) 13.63 (2.50) Corneal pachimetry (CCT) 555.58 (33.21) 529.99 (25.60)
REFERENCES
[0097] Asaoka, R. et al. (2019a) ‘Using Deep Learning and Transfer Learning to Accurately Diagnose Early-Onset Glaucoma From Macular Optical Coherence Tomography Images’, American Journal of Ophthalmology. Elsevier Inc., 198, pp. 136-145. doi: 10.1016/j.ajo.2018.10.007. [0098] Asaoka, R. et al. (2019b) ‘Using Deep Learning and Transfer Learning to Accurately Diagnose Early-Onset Glaucoma From Macular Optical Coherence Tomography Images’, American Journal of Ophthalmology. Elsevier Inc., 198, pp. 136-145. doi: 10.1016/j.ajo.2018.10.007. [0099] Banegas, Sebastián A et al. (2015) ‘Agreement among spectral-domain optical coherence tomography, standard automated perimetry, and stereophotography in the detection of glaucoma progression.’, Investigative ophthalmology & visual science, 56(2), pp. 1253-60. doi: 10.1167/iovs.14-14994. [0100] Banegas, Sebastián A. et al. (2015) ‘Agreement among spectral-domain optical coherence tomography, standard automated perimetry, and stereophotography in the detection of glaucoma progression’, Investigative Ophthalmology and Visual Science, 56(2), pp. 1253-1260. doi: [0101] 10.1167/iovs.14-14994. [0102] Blumberg, D. M. et al. (2016) ‘Technology and the glaucoma suspect’, Investigative [0103] Ophthalmology and Visual Science, 57(9), pp. OCT80-OCT85. doi: 10.1167/iovs.15-18931. [0104] Budenz, D. L. et al. (2006) ‘Detection and Prognostic Significance of Optic Disc Hemorrhages during the Ocular Hypertension Treatment Study’, Ophthalmology. doi: 10.1016/j.ophtha.2006.06.022. [0105] Chang, R. T. et al. (2009) ‘Sensitivity and specificity of time-domain versus spectral-domain optical coherence tomography in diagnosing early to moderate glaucoma.’, Ophthalmology. United States, 116(12), pp. 2294-2299. doi: 10.1016/j.ophtha.2009.06.012. [0106] Chen, Z. et al. (2019) ‘Feature Selection May Improve Deep Neural Networks For The Bioinformatics Problems.’, Bioinformatics (Oxford, England). doi: 10.1093/bioinformatics/btz763. [0107] Cordeiro, M. F. et al. (2017) ‘Real-time imaging of single neuronal cell apoptosis in patients with glaucoma’, Brain, 140(6). doi: 10.1093/brain/awx088. [0108] DeLeon Ortega, J. E. et al. (2007) ‘Effect of glaucomatous damage on repeatability of confocal scanning laser ophthalmoscope, scanning laser polarimetry, and optical coherence tomography.’, Investigative ophthalmology & visual science. United States, 48(3), pp. 1156-1163. doi: 10.1167/iovs.06-0921. [0109] Gordon, M. O. et al. (2002) ‘The Ocular Hypertension Treatment Study: Baseline factors that predict the onset of primary open-angle glaucoma’, Archives of Ophthalmology. doi: 10.1001/archopht.120.6.714. [0110] Gordon, M. O. et al. (2003) ‘Ocular hypertension treatment study: Baseline factors that predict the onset of primary open-angle glaucoma’, Evidence-Based Eye Care. doi: 10.1097/00132578-200301000-00007. [0111] Harwerth, R. S. et al. (2007) ‘The relationship between nerve fiber layer and perimetry measurements’, Invest Ophthalmol Vis Sci, 48(2), pp. 763-773. Available at: http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citatio n&list_ui ds=17251476. [0112] ‘ImageJ’ (no date). Available at: http://imagej.nih.gov/ij/.%0A. [0113] Jiang, Y. et al. (2018) ‘Optic Disc and Cup Segmentation with Blood Vessel Removal from Fundus Images for Glaucoma Detection’, in Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS. Institute of Electrical and Electronics Engineers Inc., pp. 862-865. doi: 10.1109/EMBC.2018.8512400. [0114] Jonas, J. B. et al. (2017) ‘Glaucoma’, The Lancet. doi: 10.1016/S0140-6736(17)31469-1. [0115] Kucur, Ş. S., Holló, G. and Sznitman, R. (2018) ‘A deep learning approach to automatic detection of early glaucoma from visual fields.’, PloS one, 13(11), p. e0206081. doi: 10.1371/journal.pone.0206081. [0116] Levine, R. A. et al. (2006) ‘Asymmetries and visual field summaries as predictors of glaucoma in the ocular hypertension treatment study’, Investigative Ophthalmology and Visual Science. doi: 10.1167/iovs.05-0469. [0117] MacCormick, Ian J C et al. (2019) ‘Accurate, fast, data efficient and interpretable glaucoma diagnosis with automated spatial analysis of the whole cup to disc profile.’, PloS one, 14(1), p. e0209409. doi: 10.1371/journal.pone.0209409. [0118] MacCormick, Ian J. C. et al. (2019) ‘Correction: Accurate, fast, data efficient and interpretable glaucoma diagnosis with automated spatial analysis of the whole cup to disc profile (PLoS ONE (2019) 14:1 (e0209409) DOI: 10.1371/journal.pone.0209409)’, PLoS ONE, 14(4), pp. 1-20. doi: 10.1371/journal.pone.0215056. [0119] Medeiros, F. A., Jammal, A. A. and Thompson, A. C. (2019a) ‘From Machine to Machine: An [0120] OCT-Trained Deep Learning Algorithm for Objective Quantification of Glaucomatous Damage in Fundus Photographs’, Ophthalmology. Elsevier Inc., 126(4), pp. 513-521. doi: 10.1016/j.ophtha.2018.12.033. [0121] Medeiros, F. A., Jammal, A. A. and Thompson, A. C. (2019b) ‘From Machine to Machine: An OCT-Trained Deep Learning Algorithm for Objective Quantification of Glaucomatous Damage in Fundus Photographs’, Ophthalmology. American Academy of Ophthalmology, 126(4), pp. 513-521. doi: 10.1016/j.ophtha.2018.12.033. [0122] Miglior, S. et al. (2007) ‘Intercurrent Factors Associated with the Development of Open-Angle Glaucoma in the European Glaucoma Prevention Study’, American Journal of Ophthalmology. doi: 10.1016/j.ajo.2007.04.040. [0123] Moreno-Montañés, J. et al. (2017) ‘Intraobserver and interobserver agreement of structural and functional software programs for measuring glaucoma progression’, JAMA Ophthalmology. American Medical Association, 135(4), pp. 313-319. doi: 10.1001/jamaophthalmol.2017.0017. [0124] Orlando, J. I. et al. (2019) ‘REFUGEChallenge: A Unified Framework for Evaluating Automated Methods for Glaucoma Assessment from Fundus Photographs’, Medical Image Analysis. Elsevier B V, p. 101570. doi: 10.1016/j.media.2019.101570. [0125] Pan, I., Agarwal, S. and Merck, D. (2019) ‘Generalizable Inter-Institutional Classification of Abnormal Chest Radiographs Using Efficient Convolutional Neural Networks.’, Journal of digital imaging, 32(5), pp. 888-896. doi: 10.1007/s10278-019-00180-9. [0126] Pang, S. et al. (2019) ‘An artificial intelligent diagnostic system on mobile Android terminals for cholelithiasis by lightweight convolutional neural network.’, PloS one, 14(9), p. e0221720. doi: 10.1371/journal.pone.0221720. [0127] Pathak, M., Demirel, S. and Gardiner, S. K. (2013) ‘Nonlinear, multilevel mixed-effects approach for modeling longitudinal standard automated perimetry data in glaucoma.’, Investigative [0128] ophthalmology & visual science, 54(8), pp. 5505-13. doi: 10.1167/iovs.13-12236. [0129] Poplin, R. et al. (2018) ‘Prediction of cardiovascular risk factors from retinal fundus photographs via deep learning’, Nature Biomedical Engineering. Nature Publishing Group, 2(3), pp. 158-164. doi: 10.1038/s41551-018-0195-0. [0130] Quellec, G. et al. (2017) ‘Deep image mining for diabetic retinopathy screening.’, Medical image analysis, 39, pp. 178-193. doi: 10.1016/j.media.2017.04.012. [0131] Quigley, H. A. and Broman, A. T. (2006) ‘The number of people with glaucoma worldwide in 2010 and 2020’, Br J Ophthalmol, 90(3), pp. 262-267. Available at: http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citatio n&list_ui ds=16488940. [0132] Sandler, M. et al. (2018) ‘MobileNetV2: Inverted Residuals and Linear Bottlenecks’, arXiv e- prints The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4510-4520. [0133] Schmidt-Erfurth, U., Bogunovic, H., et al. (2018) ‘Machine Learning to Analyze the Prognostic Value of Current Imaging Biomarkers in Neovascular Age-Related Macular Degeneration.’, Ophthalmology. Retina, 2(1), pp. 24-30. doi: 10.1016/j.oret.2017.03.015. [0134] Schmidt-Erfurth, U., Waldstein, S. M., et al. (2018) ‘Prediction of Individual Disease Conversion in Early AMD Using Artificial Intelligence.’, Investigative ophthalmology & visual science, 59(8), pp. 3199-3208. doi: 10.1167/iovs.18-24106. [0135] Tan, B. B. et al. (2012) ‘Comparison of retinal nerve fiber layer measurement between 2 spectral domain OCT instruments.’, Journal of glaucoma. United States, 21(4), pp. 266-273. doi: 10.1097/IJG.0b013e3182071cdd. [0136] Tatham, A. J. and Medeiros, F. A. (2017) ‘Detecting Structural Progression in Glaucoma with Optical Coherence Tomography.’, Ophthalmology. United States, 124(125), pp. S57-S65. doi: 10.1016/j.ophtha.2017.07.015. [0137] Tham, Y. C. et al. (2014) ‘Global prevalence of glaucoma and projections of glaucoma burden through 2040: A systematic review and meta-analysis’, Ophthalmology. doi: 10.1016/j.ophtha.2014.05.013. [0138] Thompson, A. C., Jammal, A. A. and Medeiros, F. A. (2019) ‘A Deep Learning Algorithm to Quantify Neuroretinal Rim Loss From Optic Disc Photographs.’, American journal of ophthalmology, 201, pp. 9-18. doi: 10.1016/j.ajo.2019.01.011. [0139] Ting, D. S. W. et al. (2019) ‘Artificial intelligence and deep learning in ophthalmology’, British Journal of Ophthalmology. BMJ Publishing Group, pp. 167-175. doi: 10.1136/bjophthalmol-2018-313173. [0140] Viswanathan, A C et al. (2003) Interobserver agreement on visual field progression in glaucoma: a comparison of methods, Br J Ophthalmol. Available at: www.bjophthalmol.com. [0141] Viswanathan, A. C. et al. (2003) ‘Interobserver agreement on visual field progression in glaucoma: A comparison of methods’, British Journal of Ophthalmology, 87(6), pp. 726-730. doi: 10.1136/bjo.87.6.726. [0142] Wang, M. et al. (2019) ‘An Artificial Intelligence Approach to Detect Visual Field Progression in Glaucoma Based on Spatial Pattern Analysis.’, Investigative ophthalmology & visual science, 60(1), pp. 365-375. doi: 10.1167/iovs.18-25568. [0143] Wang, Y. X. et al. (no date) ‘Comparison of neuroretinal rim area measurements made by the Heidelberg Retina Tomograph I and the Heidelberg Retina Tomograph II.’, Journal of glaucoma, 22(8), pp. 652-8. doi: 10.1097/IJG.0b013e318255da30. [0144] Weinreb, R. N. et al. (2018) ‘Oral Memantine for the Treatment of Glaucoma: Design and Results of 2 Randomized, Placebo-Controlled, Phase 3 Studies’, Ophthalmology, 125(12), pp. 1874-1885. doi: 10.1016/j.ophtha.2018.06.017. [0145] Weinreb, R. N. and Kaufman, P. L. (2009) ‘The glaucoma research community and FDA look to [0146] the future: a report from the NEI/FDA CDER Glaucoma Clinical Trial Design and Endpoints Symposium’, Invest Ophthalmol Vis Sci, 50(4), pp. 1497-1505. Available at: http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citatio n&list_ui ds=19321793. [0147] Weinreb, R. N. and Kaufman, P. L. (2011) ‘Glaucoma research community and FDA look to the future, II: NEI/FDA Glaucoma Clinical Trial Design and Endpoints Symposium: measures of structural change and visual function.’, Investigative ophthalmology & visual science. United States, pp. 7842-7851. doi: 10.1167/iovs.11-7895.