AUTOMATED METHOD OF IDENTIFYING A STRUCTURE
20220092824 · 2022-03-24
Inventors
Cpc classification
G01N2021/258
PHYSICS
G06V20/69
PHYSICS
G01N21/255
PHYSICS
International classification
G01N21/25
PHYSICS
Abstract
An automated method of identifying a structure in a sample is disclosed. The method includes receiving at least one digital image of a sample wherein at least one localised structural property of the sample is visible in the image based on the colour of received light. The method involves processing the at least one image, based on the received colour information to selectively identify said structure. The method can include colour and/or morphology based image analysis.
Claims
1. An automated method of identifying a structure in a sample, the method comprising: receiving at least one digital image of a sample wherein at least one localised structural property of the sample is visible in the image based on the colour of received light captured in the at least one digital image, said digital image comprising a plurality of pixels; processing the at least one image, based on the received colour information to selectively identify said structure.
2. The method as claimed in claim 1 wherein processing the at least one image includes, providing an output indicating the identification of the structure.
3. The method as claimed in any one of claim 1 or 2 which includes filtering the image to selectively process a portion of the image on the basis of colour information contained in the received image.
4. The method as claimed in any one of the preceding claims wherein the at least one image includes a plurality of pixels, and said method includes segmenting the at least one image based on a colour of received light captured in the image.
5. The method as claimed in claim 4 wherein segmenting the at least one image includes one or more of: Identifying one or more subsets of pixels within an image based at least partly on colour; Grouping pixels into features representing a structure on the sample based on correlation between a pixel and at least one neighbouring pixel.
6. The method as claimed in any one of the preceding claims which includes any one or more of: determining a colour distribution of the received image; determining a colour histogram of the received image; performing spectral analysis of at least part of the received digital image.
7. The method as claimed in any one of the preceding claims which includes performing a feature extraction method to identify one or more structures in the image.
8. The method as claimed in any one of the preceding claims wherein the method includes processing the digital image with an image recognition system.
9. The method of claim 8 wherein the image recognition system is artificial neural network.
10. The method as claimed in any one of the preceding claims wherein in the received digital image, the localised structural property of the sample is a localised refractive index.
11. The method of claim 10 wherein a structure in the sample with a given refractive index appears as a corresponding colour or colour range in the image.
12. The method as claimed in any one of the preceding claims wherein the sample is a biological sample.
13. The method as claimed in any one of the preceding claims wherein the method includes identifying a feature based on colour differentiation.
14. The method as claimed in any one of the preceding claims wherein the method includes identifying a feature based on morphology differentiation.
15. The method as claimed in any one of the preceding claims wherein the method includes identifying a feature based on a combination of colour differentiation and morphology differentiation.
16. The method as claimed in any one of the preceding claims wherein receiving at least one digital image of a sample, includes receiving more than one image of the sample captured with any one or more of the following differences: a different illumination characteristics different illumination spectrum different illumination polarisation different magnification.
17. The method of any one of the preceding claims wherein the structure includes any one or more of a neoplastic cell, cancer cell, healthy cell, cell of a given type, cell state, parasite, group of cells, abnormal cell, infected cell, tissue of a given type.
18. A method of imaging a sample and automatically generating an indication of a presence of at least one structure in the sample, the method includes: providing a sample holder having a plasmonic layer including a periodic array of sub-micron structures; placing the sample on the sample holder adjacent the plasmonic layer; illuminating the sample and sample holder and capturing at least one colour digital image thereof; performing a method of identifying a structure in the sample according to any one of the preceding claims.
19. A computer readable medium storing thereon instructions, which when executed by a data processing system cause the data processing system to perform a method as claimed in any one of claims 1 to 18.
20. A system comprising a data processing system, said data processing system being programmed to perform a method as claimed in any one of claims 1 to 18.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0030] Illustrative embodiment of the present invention will be described by way of non-limiting example with reference to the accompanying drawings. The drawings filed with the present international application include colour images used in, and arising from use of embodiments of the present invention. The colour information forms part of the disclosure of the embodiments. Should black and white or greyscale reproduction of the images occur, colour disclosure can be obtained from the originally filed documents. In the drawings:
[0031]
[0032]
[0033]
[0034]
[0035]
[0036]
[0037]
[0038]
[0039]
[0040]
[0041]
DETAILED DESCRIPTION OF THE EMBODIMENTS
[0042] The present inventors have realised that the colour contrast exhibited in the images obtained using CCM can enhanced the ability to perform automated image analysis techniques on such images, e.g. to identify one or more structure within the sample, identify one or more properties of the sample in the image.
[0043]
[0044] The present embodiment illustrates direct network connections between the elements of the system 100. The network connections may be wired or wireless, and may include a plurality of networks. In some embodiments two or more of the elements of the system 100 may not be collocated, and thus may be connected via the internet or other network adapted to enable remote communications between devices. Moreover two or more elements of the system may be combined into a single device or split into separate devices. For example the functions of the computer system 108 which operate to control the slide scanner 102 may be performed by an on-board control system computer of the slide scanner 102, or a dedicated control computer, whereas the image analysis functionality may be performed by a separate computer running suitable analysis software.
[0045]
[0046] A sample, typically a slice of tissue, which need not be stained in the preferred embodiment of the present invention, is placed on the sample holder adjacent the plasmonic layer. The sample and sample holder are loaded into the slide scanner 102. The sample scanner illuminates (at step 206) the sample and sample holder and forms as image thereof (step 208). In essence the slide scanner 102 is a microscope that illuminates the sample holder and captures an image thereof. A suitable slide scanner could for example be a member of the Aperio slide scanners from Leica Biosystems or any other slide scanner capable of capturing colour images of a microscope slide or similar sample holder. The microscope can capture images in transmission or reflection mode.
[0047] The end result of the image generation phase 202 is one or more colour images of the sample, which can then be analysed in subsequent steps of the method. In an exemplary form the samples are imaged at high resolution (e.g., 240 nm per pixel).
[0048] The images may be of the whole sample holder, essentially a “whole-slide image”. The image of the sample holder can be stored in a multi-resolution pyramid structure including a plurality of individual images providing multiple down-sampled versions of an original image. Each down sampled version is stored as a series of tiles, to facilitate rapid retrieval of sub-regions of the image. A typical whole-slide image may be approximately 200000×100000 pixels. In one form each image can be saved in RGB colour format with one byte per colour channel colour depth, although lesser or greater colour depth, or a different colour space may be used in some embodiments. The images are stored in the data storage system 106.
[0049] In the image processing phase 210 the image data of the whole slide or part of the slide are received (e.g. retrieved from data storage or received via a network connection, email etc.) and analysed to identify one or more structures (214) contained in the sample, and an output generated (216). The output can take a wide variety of forms. For example it may be, without limitation, one or more of the following:
[0050] Issuance of a notification message or alert; respect or after data reflecting the outcome of acid analysis;
[0051] Generation of an augmented image, (e.g. see
[0052] Generation of an image with metadata indicating the identification of a structure or one or more analysis steps performed thereon.
[0053]
[0054] generation of a plurality of individual images from a single image, e.g. generating multiple down-sampled versions of an original image, subdivision of an image into a series of spatial tiles.
[0055] colour correction, balancing colour response of different colour planes in the image;
[0056] noise removal or noise equalisation.
[0057] It should be noted that the pre-processing steps may be omitted if the received image data is ready for image analysis without it.
[0058] Next, the image (or multiple images together) are processed to identify one or more features contained in the image. In a primary processing step colour-based image analysis is performed. In the present invention, the colour at a particular location in the image is representative of a local physical property of the sample. In particular by using a sample holder having a plasmonic layer including a periodic array of sub-micron structures a colour contrast is exhibited in the received image. In particular, areas of the sample having different a dielectric constant appear in the image with different colours.
[0059] The underlying mechanism for the extraordinary optical contrast in the images is the resonant interaction of light with the collective oscillations of free electrons at a metal surface in the plasmonic layer of the sample holder, known as Surface Plasmon Polaritons (SPPs). The spectral change in transmitted light through an array of sub-wavelength apertures in contact with a dielectric specimen is a function of the wavelength shift, Δλ of the SPP resonant modes λ.sup.θ.sub.SPP, where superscript θ denotes the incident polarisation angle (the symbol is removed for unpolarised light) and the subscript indicates whether the dielectric constant is for the sample (d=s) or for air (d=a). The SPP modes are characterised by peaks in the transmission spectra, the corresponding wavelength shift relative to air when a sample of thickness t is placed on top of the nanoapertures is given by:
Δλ≈(λ.sup.θ.sub.SPP,s−λ.sup.θ.sub.SPP,a)(1−exp(−2t/l.sub.d)), (1)
[0060] where l.sub.d˜2√ε.sub.d is the characteristic decay length of the SPP electromagnetic field, which is itself a function of ε.sub.d, the dielectric constant of the sample. It should be noted however that in the preferred embodiments the sample is significantly thicker than the characteristic decay length of the sample. As the film thickness increases, the transmission SPP resonance peak is increasingly red-shifted until it equals λ.sup.θ.sub.SPP, after which no more colour change occurs. It follows that, when using a standard transmission bright-field microscope, or reflection microscope, a spatially resolved distribution of colours will result that relates directly to changes in the local dielectric constant in the sample. With the local dielectric constant encoded in the optical spectrum, a remarkable chromatic contrast effect is produced. This means that any structure within optically transparent samples, which previously was difficult to detect due to a lack of contrast, is detectable in the visible-light image, by virtue of the colour contrast captured in the images.
[0061] Next in step 222 an image (or two or more images) is analysed to identify features in the image (or two or more images) that are representative of one or more structures of interest in the sample. A structure of interest can, for example include, a cell, group of cells, part of a cell, interstitial space between cells, void in a cell, the morphology of any of the above. In some embodiments the method can include any one or more of the following processing steps or sub-steps: [0062] Colour filtering the image to selectively process (e.g. perform feature differentiation) on one or more colour bands of the received image, or an image generated from one or more colour bands; [0063] Determining a colour distribution or colour histogram of the received image; [0064] Performing a feature extraction method to identify one or more structures in the image; [0065] Processing the digital image with an image recognition system. [0066] The analysis in step 222 can be grouped into two general types of processes: [0067] Colour differentiation (224), which encompasses differentiation methods based on the localised colour of the image of the sample, relative colour of more than two locations in the image, overall colour/spectral content of the image or portion thereof. Colour differentiation can be based on a CIE plot which mimics sensitivity of the human eye. In some embodiments a spectral ‘fingerprint’ can be generated for an image or portion thereof. The spectral fingerprint can enable either a coarse analysis based on e.g. where the spectral fingerprint in the form of an RGB ratio is compared to values in a look up table to determine if a sample is ‘healthy’ or not. Alternatively a more detailed spectral fingerprint can be compared against a higher resolution spectrum e.g. collected using a spectrometer. As will be appreciated in analysing any one image (or set of two or more images) multiple colour differentiation techniques can be used. An exemplary colour-based analysis method 222 is described below in relation to
[0069] An exemplary colour-based differentiation process 222 is shown in
[0070] Both methods generally follow the process of Shi, P. et al. Automated Ki-67 Quantification of Immunohistochemical Staining Image of Human Nasopharyngeal Carcinoma Xenografts. Scientific Reports 6, 32127, doi:10.1038/srep32127 (2016), the contents of which are incorporated herein by reference.
[0071] For both the nanoslide images and Ki67 images the mean RGB space and HSL space values for the cancer cells were determined from the ground truth standard. Cancer cells when imaged on the nanoslide manifest themselves as generally blue in hue, whereas, Ki-67 positive nuclei manifest themselves as brown hue in images of breast tissues.
[0072] The mean RGB and HSL channel values for positive cancer cells in Ki67 and nanoslide are summarised in Table 1 below. The RGB values for Ki67 positivity determined by the inventors are close to the published values from (Shi et al., Scientific Reports, 2016).
TABLE-US-00001 TABLE 1 Mean RGB Mean HSL space values space values Values R G B H S L Ki67 (brown) 123 51 7 23 89 26 Nanoslide (blue) 23 69 86 196 58 21
[0073] Based on the variability of the colour change associated with cell positivity in nanoslide and Ki67 a ±15% threshold centred around the mean HSL colour space values (in each of H, S and L) was used for segmentation of positive cancer cells—that is, within this range cells were considered to be ‘positive’ for cancer. An example range of HSL colour space values corresponding to cancer positivity using nanoslide is shown in
[0074] In the local feature extraction step 504 pixels of the image(s) having HSL values corresponding with cancer positivity are selected. In the present example using the nanoslide only a single colour range is considered, but other embodiments may perform multiple segmentations based on different colour ranges to attempt to identify respective structures in the image(s). Pixels having colour values in the selected range are further analysed to identify groups or subsets of pixels that may represent the structure of interest (e.g. nucleus of a cancer cell). The subsequent analysis can include analyzing the pixels to identify correlation between a pixel and at least one neighbouring pixel. In some examples, the mean intensity and standard deviation at a pixel site can be calculated based on a 3×3 group of pixels centred on the pixel, and can be used as a measure of signal strength and local variance on signal strength in the neighbourhood of the pixel site. The skewness and kurtosis of pixel and its surrounding 3×3 neighbourhood of pixels can be calculated. The skewness measures symmetry of local intensity and whether the intensity is more peaked or flat than a normal distribution. These measures of local pixel correlation can be used to cluster similar pixels or separate dis-similar pixels into clusters (representing structures of interest) in step 508. Grouping can be performed using kmeans clustering or other suitable techniques, as part of a structure identification step 506. It may be necessary to more accurately define the boundary of a structure of interest. In step 510, clusters representing touching or overlapping features can be separated into individual clusters. This can be performed, e.g. using the watershed method.
[0075] In step 228 the output of the differentiation processes are used to make a decision on the presence or absence of one or more features in the image which indicate the presence or absence of a structure in the sample.
[0076] In some embodiments a combination of colour differentiation in addition to sample morphology differentiation enables more accurate identification of a structure. For example a spatial correlation between the output form colour differentiation and morphology differentiation could be used to distinguish between features that are indistinguishable by one process alone.
[0077] The image recognition system can be an artificial neural network. The image recognition system can be trained on a suitable training set of images containing known structures. Saha, M. et al., An exemplary image recognition method and system that could be used in embodiments of the present invention is described in “An Advanced Deep Learning Approach for Ki-67 Stained Hotspot Detection and Proliferation Rate
[0078] Scoring for Prognostic Evaluation of Breast Cancer. Scientific Reports; 7: 3213 DOI:10.1038/s41598-017-03405-5, (2017) describes a deep learning methodology using an ANN system to detect breast cancer cells for Ki67 stained samples. Such a methodology could be applied to images taken with a nanoslide to identify similar structures as taught herein.
[0079] In the event that one or more features are identified in step 228 the one or more images, an output is generated at 232. The output can take many forms. The output can take a wide variety of forms. For example it may be, without limitation, one or more of the following:
[0080] Issuance of a notification message or alert;
[0081] Generation of an augmented image (in step 230), e.g. on which an identified structure or candidate feature which may be an identified structure is enhanced e.g. it can be emphasised, rendered more visible or rendered identifiable to a user;
[0082] Generation of an image with metadata indicating the identification of a structure or one or more analysis steps performed thereon.
[0083] Feature enhancement can include showing the boundary of the structure or structures which have been identified in the image(s) in an augmented image. The boundary can include drawing a bounding box or other outline around the identified features. This can be of use to a user of the system to rapidly enable identification and human investigation and verification of the presence or absence of the structure.
[0084] Methods disclosed herein can be used in an implementation an embodiment of an aspect of the applicant's co-pending Australian patent application 2018904550, filed on 29 Nov. 2018, entitled “Method of identifying a structure” and the International patent application claiming priority to AU 2018904550 which was filed on the same day as present application.
EXAMPLE
[0085] To illustrate the usefulness of the present method, automated image processing is performed using nanoslide images on the basis of the HSL (hue, saturation, lightness) of the image pixels. The results were compared to those obtained from analysis of samples stained using proliferative markers (Ki67) and illustrate the technique can distinguish a structure, which in this example is distinguishing neoplastic cells, as compared to normal breast epithelium.
[0086] In the present example sample holders having a plasmonic layer with a periodic array of sub-micron structures (e.g. circular holes) were fabricated using displacement Talbot lithography to produce devices on the scale of whole microscope slides (e.g. 75 mm×25 mm). The nanoslides were used in conventional brightfield imaging of samples and histological tissue sections of corresponding samples were scored by two reference pathologists. The properties of the brightfield images were analysed based on the measured hue, saturation, and lightness in the image.
[0087]
[0088] In the study the images made use of the MMTV-PyMT model of spontaneous breast tumorigenesis, where mice develop pre-invasive and invasive neoplasms within 50 days of age. Pre-invasive and invasive neoplasms have previously been shown to be distinguishable from benign epithelial cells using IHC staining for the proliferative marker Ki67. In total 24 mice were used for this study. The workflow for the study design is shown in
[0089] For the nanoslide and Ki67 stained samples automated image analysis as set out above was performed in step 700 to identify the presence or absence of neoplastic cells.
[0090]
[0091] Following established protocols tissue (as identified by the automated image analysis of the nanoslide images) was classified as True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) as compared to human analysis. In the human classification, pathology annotations, when a cancer containing region has been identified, high-resolution H&E stained slides were used to identify the stage of the cancer and the margins. A morphological assessment of the tissues was conducted by an expert human breast and murine mammary gland pathologist (O'Toole) and breast cancer researcher (Parker) and formed the ‘ground truth’ for the analysis. The second piece of information came from the image pixel HSL colour space values which were compared against the reference values from the training data (as set out for the nanoslide in
TABLE-US-00002 Description of classification method for Ki67 Classification and Nanoslide True TP was assigned when the HSL colour space values Positive were consistent with cancer cells established by (TP) ‘training’ the segmentation algorithm. This ‘training’ was conducted based on the identification and correlation of cancerous tissue in Ki67 and nanoslide images by the expert pathologist with reference to the H&E slides (e.g. Shi et al, Scientific Reports, 2016). To be classified as TP also required that the identified region was within the area manually identified as containing cancer cells by the expert pathologists. True TN was assigned when the HSL colour space Negative values were consistent with one of the sub-types of (TN) non-cancerous tissues (e.g. adipose tissue, collagen, lymph nodes, blood vessels etc.). To be classified as TN also required that the identified region was outside of the area manually identified as containing cancerous tissue by the expert pathologists. False FP was assigned when the HSL colour space Positive values were consistent with cancer cells (FP) but the identified region was outside of the area manually identified as containing cancer cells by the expert pathologists. False FN was assigned when the HSL colour space Negative values were not consistent with either cancer cells or (FN) with non-cancerous tissue and when the identified region was within the area manually identified as containing cancer cells by the expert pathologists.
[0092]
[0093] The methods disclosed herein use utilise the differences in the spectral output between structures to identify those structures.
[0094]
[0095] To test the concordance of Ki67 and nanoslide we compared the percentage (by area) of tissue identified by the two pathologists as containing neoplastic cells according to the image pixel HSL colour space values; the results are summarised in
DSC=2TP/(2TP+FP+FN)
[0096] Calculated for both nanoslide and Ki67 (
[0097]
[0098] This compares favourably in to IHC staining based on proliferative markers (
[0099] Pathology Assessment
[0100] In the example to confirm the timing of spontaneous development of mammary gland tumours in the C57 Bl/6 MMTV-PyMT model, mammary glands of C57 Bl/6 MMTV-PyMT mice at different stages were taken and morphologically evaluated by H&E and Ki67 by an expert human breast and murine mammary gland pathologist (O'Toole) and breast cancer researcher (Parker). Nanoslide samples were randomized and independently scored and then compared post-analysis to the results of Ki67 and nanoslide. The benchmark for the pathology assessment was a trained pathologist analyzing the H&E stained tissue sections at high-resolution and without any time constraints. As this was a control study the cancer stage for the mice was already known by the pathologist. In addition, the pathologist could refer back to the IHC staining to confirm that no neoplastic tissue regions were missed during the assessment. When looking at a tumor region or duct containing cancer at high resolution the pathologist counts the number of cancer cells.
[0101] Once this has been done for all samples the pathologist then compared the number of individual positive cells (as determined by a colour change—‘brown’ for Ki67 and ‘green’ for nanoslide) using either Ki67 or nanoslide and divided this number by the total number of cancer cells identified from pathological assessment of the H&E images to arrive at the final figure for “percentage positive cells”. This analysis was conducted on 24 cancer containing regions across the 24 mice used in this study. Based on the knowledge of the cancer stage the results could be classified into 4 stages: ‘normal’, ‘hyperplasia’, DCIS′, and ‘invasive’. The mean value of the percentage of positive cancer cells as determined by the pathologist was calculated within each category, it is this mean value, averaged between the two independent sets of scores, which is represented by the height of the bars in the bar chart. The range (e.g. minimum and maximum percentages) over the different samples used to generate the error bars shown in
TABLE-US-00003 Normal DCIS Invasive Appearance of lumen Empty Lumen Filled Lumen No Lumen Epithelial Ki67 positivity 0-28% 44-66% 48-96% (95% confidence interval)
[0102] It will be understood that the invention disclosed and defined in this specification extends to all alternative combinations of two or more of the individual features mentioned or evident from the text or drawings. All of these different combinations constitute various alternative aspects of the invention.