SYSTEM AND METHOD FOR DIGITALIZATION, ANALYSIS AND STORAGE OF BIOLOGICAL SAMPLES
20220367013 · 2022-11-17
Assignee
Inventors
Cpc classification
G16H10/40
PHYSICS
G02B7/36
PHYSICS
G16H40/20
PHYSICS
G16H50/70
PHYSICS
G16H50/20
PHYSICS
G06F21/604
PHYSICS
G01N35/00871
PHYSICS
International classification
G16H10/40
PHYSICS
G01N35/00
PHYSICS
Abstract
Biological samples are prepared on a slide for physician or veterinarian interpretation in a case of a diagnosis for human or animal diseases. The biological samples are specimens taken from certain areas or body fluids of human or animal. The biological samples are placed on the slide and are made ready for an examination without any process or after physical processes. Physicians or veterinarians diagnose by interpreting the biological samples over a microscope. Digitizing data, usage of data processing techniques and automatic reporting are activities reducing a workforce of expert physicians or veterinarians with a developing technology by abandoning manual methods. As digital pathology and hematology are main technical fields, a system and integrated methods about digitizing, analyzing and storing biological samples are given.
Claims
1. A method for enabling biological samples on slides to be automatically scanned and analyzed by algorithms on a cloud, comprising following steps of: a) pairing hardware and the cloud using an Internet of Things (IOT) password; b) placing the slides on a slide plane; c) focusing; d) getting a sampling number and an analysis type from the cloud; e) analyzing the biological samples on the cloud; f) collection of image data; and g) hierarchical storage of the image data.
2. The method according to claim 1, wherein after step a), the slides are encoded with barcodes output from the cloud.
3. The method according to claim 2, wherein the barcodes are printed from the cloud for patients registered in the cloud.
4. The method according to claim 1, wherein the analysis type is determined manually or by a barcode.
5. The method according to claim 1, wherein an immersion oil is dripped onto the slides.
6. The method according to claim 1, wherein 10×, 40× and 100× magnification lenses are automatically selected according to the analysis type.
7. The method according to claim 1, wherein single or multiple images are collected according to the analysis type.
8. The method according to claim 1, wherein the step of focusing is manual or automatic.
9. The method according to claim 7, characterized in that it includes wherein a monitor, a keyboard and a mouse are configured for a user-controlled manual focusing.
10. The method according to claim 7, wherein a set of subjects is prepared in advance for an automatic focusing, to train a system definitions of active and noisy zones and to focus automatically at one time without a need for a learning during the step of focusing.
11. The method according to claim 1, wherein the image data is received by a user via the hardware or manually.
12. The method according to claim 11, characterized in that comprising the following steps when the image data is received via the hardware; a) sending a biological sample image and a data processing type to an encryption and data decoder structure of a cloud system via the hardware; b) redirecting, by the encryption and data decoder structure, the biological sample image to an algorithm adapter web service according to the data processing type; c) running, by the algorithm adapter web service, a data processing method for the biological sample image and retrieving results from an algorithm block; d) sending, by the encryption and data decoder structure, the results to the hardware; e) sending, by the hardware, aggregate results to the encryption and data decoder structure together with barcode data; f) sending by the encryption d data decoder structure, the results to a user approval via an internet interface; g) opening the results to a user access on the cloud after the user approval.
13. The method according to claim 11, comprising the following steps when the image data is received manually by the user: a) logging in to the cloud; b) selecting a data processing type and a patient to associate; c) uploading a biological sample image to a cloud system; d) sending the biological sample image and the data processing type of the cloud system to an encryption and data decoder structure via an internet interface; e) redirecting, by the encryption and data decoder structure, the biological sample image to an algorithm adapter web interface according to the data processing type; f) running a data processing method for a relevant biological sample image by the algorithm adapter web interface and retrieving results from an algorithm block; g) sending, by the encryption and data decoder structure, the results to the algorithm adapter web interface for the relevant biological sample image; h) sending, by the encryption and data decoder structure, the results for a user approval via the internet interface; and i) opening the results to a user access on the cloud after the user approval.
14. The method according to claim 1, wherein a property of the analysis type is a peripheral blood smear analysis, a bone marrow analysis, an analysis of a Thoma slide, a nosema illness detection at bee samples, or a lymph node analysis.
15. A system enabling automatically scanning and analyzing of the biological samples on the slide by the algorithms on the cloud, wherein the system is running with the method according to claim 1.
Description
FIGURES THAT ARE HELPFUL TO UNDERSTAND THE INVENTION
[0031]
[0032]
[0033]
[0034]
[0035]
[0036]
[0037]
[0038]
[0039]
[0040]
[0041]
[0042]
[0043]
[0044]
[0045]
[0046]
DETAILED DESCRIPTION OF THE INVENTION
[0047] The system consists of hardware and software components.
[0048] The hardware component includes XYZ axis motors, motor controller, microscope light, zeroing switches, CCD camera, ocular, immersion oil dripping system, automatic lens changer, and Internet of Things (IoT) microprocessor (C1.13).
[0049] The hardware is connected to the cloud software with the IoT password provided from the user accounts registered in the cloud. The method to obtain this password is described in the Cloud-C2 section.
[0050] The software component consists of 3 different software elements that communicate with each other.
[0051] The first software element is an embedded software built on the Internet of Things (IoT) microprocessor (C1.13), which is used to automate the data processing process.
[0052] The second software element is an item to enable the user to review the images acquired by the hardware and it is installed to provide an interface between the hardware and the user over the cloud.
[0053] At the same time, this software is a web page that allows users to register and opens test reports coming from the hardware devices which are associated within the same institution to the registered users. This interface is described in detail in the Cloud-C2.
[0054] The third software element is an algorithm service that includes an artificial intelligence algorithm which takes an image as input and gives the results of the artificial intelligence analysis. This interface is described in detail in the Cloud-C2.
[0055] The system is designed to be cost-effective to reduce costs for the examination of patient or animal samples in small budget health care facilities and accelerate the process. The designs of the present invention are considered according to the cost-effective system and should be considered in this context.
[0056] The system can be divided into two components, C1 (hardware for digitizing the biological samples) and C2 (cloud for data analysis). The general system level block diagram is given in
[0057] 4.1 Hardware—C1
[0058] The diagram which shows the hardware components inside C1 can be seen in
[0059] Detailed technical drawings (C1) of the equipment are shown in
[0060] C1.13 is the main control unit for all hardware components in the hierarchy to digitize the biological sample.
[0061] The inputs of the hardware system are the biological sample, AC power, immersion oil tank and starting switch.
[0062] The data interface between the system and the cloud is wired or wireless Ethernet.
[0063] A barcode can be generated for the relevant human or animal from the interface on the cloud with the patients recorded over the cloud. This barcode allows the hardware to digitize and send biological samples to the cloud without knowing any specific information about humans or animals.
[0064] After inserting the biological sample with the printed barcode affixed on it to the device or manually scanning the barcode under the C1.1 Barcode scanner and placing the biological sample, the data collection process starts by pressing the ‘Start’ button.
[0065] The biological sample inspection procedure depends on the type of sample.
[0066] The type of sample is used to decide on the magnification levels. The system has different magnification levels such as 10×, 40× and 100×. For example; a minimum of 10× magnification and a single image is sufficient for the analysis of Thoma slides. However, peripheral blood smear (PBS) slides require a magnification level of 100× with immersion oils to analyze blood cell disorders and multiple images. The first software element is involved at this stage. The first software item that reads the barcode sends the barcode information to the Cloud as defined in
The Y-axis in the flow is given as the Z-axis in
[0092] 4.1.1 Auto Focus Method
[0093] The automatic focus of biological samples in a cost-effective system without position feedback for the Z axis on the magnification levels of 10×, 40× and 100× is a problem that is studied by engineers. The system, which is designed for laboratories and small health institutions, has to have the ability to perform autofocus without user interaction at all magnification levels, since it is a necessary feature for the entire steps.
[0094] The invention proposes a method by changing the lens position on the Z axis and performing auto focus, by controlling the camera outputs C1.4. This method is contained in the software block installed in the biological sample digitization hardware (C1). The system does not have a position feedback which shows whether the motor controller applies movement order given by C1.13 or not in 5 μm precision for the Z axes by considering the cost-effective system has a 5 μm precision.
[0095] In case of small position changes, system control cannot be achieved in the specified direction as desired. However, the system can control the position on the Z-axis at a position movement greater than 200±5 μm. The errors in this stage are not regarded.
[0096] The procedure, scans the aforementioned region between Z.sub.l and Z.sub.h in 5 μm slices twice times, the reality accepted at this stage is defined with the expression Z.sub.h≥Z.sub.l+200 μm. The first scan is used to extract noise and active zone Gaussian distributions. The second scan is used to focus the sample using the gauss distributions calculated in the first scan.
[0097] The procedure, defines the first scan to find K=2 number of Gaussians, divides the region into 5 μm slices with the number of N>40 samples. The scanning iterations can be seen from
[0103] To find the centroids of Gaussian clusters, the data is divided into K=2 for noise and active zone. The procedure is: [0104] (a) K points are randomly selected from the data set. They are defined as cluster centers. [0105] (b) The distance between each element of the data and the centers of the previously defined clusters is found. For example, let's take two sampling points F.sub.1 and F.sub.2, the distance between these two points is calculated as follows:
D=√{square root over (F.sub.1.sup.2−F.sub.2.sup.2)} (4) [0106] (c) Sampling points are assigned to the nearest centroid according to the distance formula. [0107] (d) The average value of each cluster is calculated by averaging the samples associated with each cluster. [0108] (e) Items between b-d are repeated until the convergence point. [0109] The clustered regions contain a data set that is distinguished from each other as in
[0112] The following procedure describes the conditions required to find the focusing position during the second scan using a trained structure by the data set obtained in the first scan. [0113] Focusing procedure starts from Z.sub.l to repeat in 5 μm slices. [0114] The focus parameter (F.sub.i) is associated with noise (G.sub.n), active (G.sub.a), and peak regions (P) according to the following conditions. When condition P is fulfilled, the system is in the autofocus position condition.
F.sub.i≤μ.sub.n+3*σ.sub.n,G.sub.n (7)
μ.sub.n+3*σn≤F.sub.i≤μ.sub.a+3*σ.sub.a,G.sub.a (8)
μ.sub.a+3*σ.sub.a≤F.sub.i,P (9) [0115] The system stops at that position and C1.3 CCD camera is used to capture image. [0116] The procedures after this step are the operations for the acquisition of the desired number of images in the horizontal plane and uploading the images to the cloud.
[0117] 4.2 Cloud—C2
[0118] Cloud computing is a powerful tool of choice with the advantages of rapid adaptation for the algorithms and the ability to connect the hardware via the Internet over the cloud that have a physical IP and address.
[0119] Installing data processing methods on each hardware causes cost-increasing results and makes software updates difficult. Therefore, this invention has taken the burden of data processing methods from the hardware and draws it to a central area that all connected hardware can use. In this way, hardware can be produced in a cost-effective manner and also software updates can be made fast and a safe working environment can be provided.
[0120] The cloud system in the mentioned invention, connects the slide scanner hardwares that are used to scan the biological samples to the central system. In this way, data coming from hardware installed in different environments can be synthesized and archived.
[0121] There are two software elements in the cloud system. The first one includes the C2.1 Web interface and the C2.2 encryption and data decoder structure, and the other one is the software component called as C2.3 algorithm adapter web service, which plays a role in system administration. The processes in the cloud (C2) object are examined in
[0122] Hardware-Acquired Images: [0123] 1) The biological sample image and data processing type are sent to the cloud system “Encryption and data decoder structure” with the hardware C1. [0124] 2) C2.2 “Encryption and data decoder structure” directs the biological sample image to C2.3 “Algorithm Adapter Web Service” according to the data processing type. [0125] 3) C2.3 “Algorithm Adapter Web Service” runs the data processing method for the relevant biological sample image and the results are taken directly from the algorithm block. [0126] 4) C2.2 “Encryption and data decoder structure” sends the results to the hardware for the corresponding biological sample image. [0127] 5) Hardware (C1) repeats steps 1-4 for each biological sample image. When data processing of all images is finished, the hardware (C1) sends the batch results with the barcode data together to C2.2 “Encryption and data decoder structure”. The data flow can be followed in
[0130] Images Taken Manually by the User: [0131] 1) The user enters the Cloud (C2) with a password and a user name. [0132] 2) The user selects the data processing type and the patient to associate. [0133] 3) The user uploads the biological sample images to the system via a web page. [0134] 4) The biological sample image and the data processing type are sent to the cloud system C2.2 “Encryption and data decoder structure” via C1.2 the Web Interface. [0135] 5) C2.2 “Encryption and data decoder structure” directs the biological sample image to C2.3 “Algorithm Adapter Web Interface” according to the data processing type. [0136] 6) C2.3 “Algorithm Adapter Web Service” runs the data processing method for the relevant biological sample image and the results are taken directly from the algorithm block. [0137] 7) C2.2 “Encryption and data decoder structure” sends the results to the C2.1 “Web Interface” for the corresponding biological sample image. [0138] 8) C2.1 “Web Interface” repeats steps 4-7 for each biological sample image. When all images are finished, it sends the batch results to the C2.2 “Encryption and data decoder structure” with the associated patient information. [0139] 9) C2.2 “Encryption and data decoder structure” sends all collected results to user approval in C2.1 “Web Interface”. [0140] 10)After user approval, the data processing results of the batch images are made available to the user on the cloud (C2).
[0141] As these steps are applied for digitizing and data processing of biological samples; different data processing methods can be used for different types of biological samples. As an example; A data processing method installed on the system for peripheral blood smear samples is described below.
[0142] 4.2.1 Example of Biological Sample Data Processing Algorithm
[0143] Peripheral Blood Smear is frequently used test by hematology and pathology departments for the diagnosis leukemia, anemia and thalassemia with the help of expert physician. This test is used in the form of one drop of blood taken from the person on the slide, after smearing, staining and washing, and the diagnosis is made by the expert physicians after the examination under the microscope.
[0144] The invention proposes a method for digitizing biological samples, which provides a peripheral blood smear data processing method as a sample algorithm data processing method. This method uses 3 different blocks for data processing of the acquired biological sample images. These blocks are segmentation, identification and correcting blocks. These blocks are shown in
[0145] Separation block can be examined with 6 different sub-components. These sub-components work for a 2-dimensional 3-color image; gray scale conversion, thresholding, edge detection, euclidean distance calculation, local maximum points and watershed algorithms. These sub-components are shown in
[0148] The output of the segmentation block results the regions defined for each cell. These regions refer to an object but there is no information about the type of the cell.
[0149] The identification block contains a trained artificial intelligence. This artificial intelligence has extracted the features of the images for each block for 20000 different cell samples, by using convolutional neural network (CNN) blocks. The last value indicates the cell type. These steps were introduced to the system by expert physicians.
[0150] For the object identification in the defined region, the result of the artificial intelligence is obtained. The block of artificial neural networks used for this step is described in
[0151] The correction block is a result correction process for the output of segmentation and identification blocks. For each cell region, Jaccard index (JI) is used to test its relationship with other regions. JI can be expressed as follows. For example, the JI value between zone A and zone B is calculated as follows.
[0152] Correction is done as in
[0153] As this part is an example of an algorithm developed for peripheral blood smear, the algorithm blocks given in
[0154] 4.2.2 Object-Based Storage of Biological Sample Images
[0155] The presented invention comprises structures that store biological samples in the cloud in a particular hierarchy and facilitate data analysis.
[0156] The storage of biological samples is organized by the hardware with the following preliminary information. Section 4.2.3 describes these steps. [0157] IoT Password that enables pairing with the cloud. [0158] The barcode on the biological sample is an output generated by the cloud. This barcode can be pasted onto the sample and the following informations can be extracted from the barcode by the cloud. [0159] Hospital ID [0160] Patient ID [0161] Analysis Type
[0162] The IoT password is unique to the physician registered on the system. As this password is given to the user over the cloud, the hardware is initialized with the help of this password and the corresponding hardware is associated with the cloud. The analysis steps are shown in
[0170] 4.2.3 IoT Password and Barcode Encryption Method
[0171] There is a general purpose private key on the system. This private key is a valid password for the entire system. This switch is only held on C2 (Cloud).
[0172] One-way encryption infrastructure has been created for encryption, and personal, hospital or physician-specific information is only kept on the cloud. The hardware communicates with the cloud through passwords. Person, hospital or physician-specific information is not held by the hardware.
[0173] IoT Password information is generated using the following information. [0174] Hospital ID [0175] Doctor ID [0176] Private Key
[0177] The cloud tests the accuracy of the password by comparing the IoT Password information sent to it with different combinations of hospitals and physicians. It continues with a correct password.
[0178] Barcode information is generated using the following information. [0179] Hospital ID [0180] Patient ID [0181] Analysis Type [0182] Private Key
[0183] The cloud tests the barcode correctness by comparing the incoming barcode information with different combinations of hospitals, patients and analysis types. If it is a correct barcode, patient number, hospital number and analysis type information will be used during the analysis process. Analysis type information is sent to the hardware for lens initialization by the cloud.