Tear film break-up time measurement for screening dry eye disease by deep convolutional neural network
10779725 ยท 2020-09-22
Assignee
Inventors
Cpc classification
A61B3/0025
HUMAN NECESSITIES
G16H50/20
PHYSICS
International classification
A61B3/10
HUMAN NECESSITIES
A61B3/14
HUMAN NECESSITIES
Abstract
A convolutional neural network model distinguishes eyelash images, break-up area images, non-break-up images, sclera images and eyelid images corresponding to a first prediction score, a second prediction score, a third prediction score, a fourth prediction score and a fifth prediction score to respectively produce a first label, a second label, a third label, a fourth label and a fifth label, thereby a break-up area can be detected in a tear film image and a tear film break-up time can be quantized for detection.
Claims
1. A tear film break-up time measurement method for screening dry eye disease by deep convolutional neural network, comprising steps as following: providing a convolutional neural network model including a first convolutional layer, a max pooling layer, a second convolutional layer, a first average pooling layer, a third convolutional layer, a second average pooling layer, a fourth convolutional layer, a third average pooling layer, a fifth convolutional layer, a fully connected layer and a softmax layer; setting sizes of a first filter, a second filter, a third filter, a fourth filter, a fifth filter and a sixth filter and disposing respectively on said first convolutional layer, said second convolutional layer, said third convolutional layer, said fourth convolutional layer, said fifth convolutional layer and said fully connected layer; dividing and selecting a plurality of eyelash images, a plurality of break-up area images, a plurality of non-break-up area images, a plurality of sclera images and a plurality of eyelid images in a tear film image to said first convolutional layer, said max pooling layer, said second convolutional layer, said first average pooling layer, said third convolutional layer, said second average pooling layer, said fourth convolutional layer, said third average pooling layer, said fifth convolutional layer, said fully connected layer and said softmax layer; forming said eyelash images, said break-up area images, said non-break-up area images, said sclera images and said eyelid images to produce a first feature map through said first filter, said first feature map being processed by said max pooling layer, and then producing a second feature map through said second filter, said second feature map being processed by said first average pooling layer, then producing a third feature map through said third filter, said third feature map being processed by said second average pooling layer, then producing a fourth feature map through a fourth filter, said fourth feature map being processed by said third average pooling layer to said fully connected layer, so that results of the eyelash images, the break-up area images, the non-break-up area images, the sclera images and the eyelid images are inserted to the fully connected layer; classifying said eyelash images, said break-up area images, said non-break-up area images, said sclera images and said eyelid images output from the fully connected layer through said softmax layer, said eyelash images, said break-up area images, said non-break-up area images, said sclera images and said eyelid images respectively corresponding to a first prediction score, a second prediction score, a third prediction score, a fourth prediction score and a fifth prediction score to respectively produce a first label, a second label, a third label, a fourth label and a fifth label; and distinguishing said eyelash images, said break-up area images, said non-break-up area images, said sclera images and said eyelid images by said first label, second label, third label, fourth label and said fifth label to detect a tear film break-up area in said tear film image as the image changes with time passing by and to quantize a tear film break-up time.
2. The tear film break-up time measurement method as claimed in claim 1, wherein the tear film break-up time is set at 5 seconds.
3. The tear film break-up time measurement method as claimed in claim 2, wherein the feature maps are either stable or unstable.
4. The tear film break-up time measurement method as claimed in claim 1, wherein the tear film image is fluorescent, ultraviolet, visible lighting, infrared or thermal.
5. The tear film break-up time measurement method as claimed in claim 1, wherein the first filter has a number of 32, a 55 size and a stride of 1, and after max pooling in the max pooling layer, the size of the first filter is altered to 33 and the stride is altered to 2; the second filter has a number of 32, a 55 size and a stride of 1, and after average pooling in the first average pooling layer, the size of the second filter is altered to 33 and the stride is altered to 2; the third filter has a number of 64, a 55 size and a stride of 1, and after average pooling in the second average pooling layer, the size of the third filter is altered to 33 and the stride is altered to 2; the fourth filter has a number of 64, a size of 55 and a stride of 1; the fifth filter has a number of 128, a size of 44 and a stride of 1; and the sixth filter has a number of 3, a size of 11 and a stride of 1.
6. The tear film break-up time measurement method as claimed in claim 5, wherein the tear film image is divided into a grid as each segment of the grid corresponding to one of the eyelash images, break-up area images, non-break-up area images, sclera images and eyelid images with 9696 pixels of each segment; a first region of interest retrieved from one of the segments with 9696 pixels for detecting the break-up area images and then extended to a second region of interest with 260260 pixels, said second region of interest having a third region of interest with 9696 pixels and a stride of 20, said third region of interest connected to the convolutional neural network model for segmentation of the second region of interest with 260260 pixels and output of a probability of said second region of interest being a break-up area image, said probability then added to a probability map, said probability map set selecting a fourth region of interest with an average probability exceeding 0.8 and a center of said fourth region of interest defined as a center of break-up, said fourth region of interest thereby creating a fifth region of interest for detection of the tear film break-up area.
7. The tear film break-up time measurement method as claimed in claim 1, wherein the tear film image is detected for eyes opening and closing by a distance between the eyelids in the image before dividing and selecting the eyelash images, the break-up area images, the non-break-up area images, the sclera images and the eyelid images in the tear film image.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
(12) Referring to the flow diagram in
(13) Step 1 S1: providing a convolutional neural network model 10 including a first convolutional layer 11, a max pooling layer 12, a second convolutional layer 13, a first average pooling layer 14, a third convolutional layer 15, a second average pooling layer 16, a fourth convolutional layer 17, a third average pooling layer 18, a fifth convolutional layer 19, a fully connected layer F and a softmax layer T.
(14) Step 2 S2: setting sizes of a first filter 111, a second filter 131, a third filter 151, a fourth filter 171, a fifth filter 191 and a sixth filter F.sub.1; the filters are respectively disposed on the first convolutional layer 11, the second convolutional layer 13, the third convolutional layer 15, the fourth convolutional layer 17, the fifth convolutional layer 19 and the fully connected layer F. Referring to
(15) Step 3 S3: dividing and selecting a plurality of eyelash images 21, a plurality of break-up area images 22, a plurality of non-break-up area images 23, a plurality of sclera images 24, and a plurality of eyelid images 25 in a tear film image V to the first convolutional layer 11, the max pooling layer 12, the second convolutional layer 13, the first average pooling layer 14, the third convolutional layer 15, the second average pooling layer 16, the fourth convolutional layer 17, the third average pooling layer 18, the fifth convolutional layer 19, the fully connected layer F and the softmax layer T. In this embodiment, the partial image 20 is one of the eyelash images 21, break-up area images 22, non-break-up area images 23, sclera images 24 and the eyelid images 25.
(16) Before dividing and selecting the eyelash images 21, the break-up area images 22, the non-break-up area images 23, the sclera images 24 and the eyelid images 25 in the tear film image V, the tear film image V is detected for eyes opening and closing by a distance between the eyelids in the image. In
(17) Step 4 S4: forming the eyelash images 21 the break-up area images 22, the non-break-up area images 23, the sclera images 24 and the eyelid images 25 to produce a first feature map M.sub.1 through the first filter 111. The first feature map M.sub.1 is then processed by the max pooling layer 12, producing a second feature map M.sub.2 through the second filter 131. The second feature map M.sub.2 is processed by the first average pooling layer 14, producing a third feature map M.sub.3 through the third filter 151. The third feature map M.sub.3 is processed by the second average pooling layer 16, producing a fourth feature map 171 through a fourth filter 171 The fourth feature map M.sub.4 is processed by the third average pooling layer 17 to the fully connected layer F, so that the results of the eyelash images 21, the break-up area images 22, the non-break-up area images 23, the sclera images 24 and the eyelid images 25 are inserted to the fully connected layer F. In short, the features can capture the discriminatory information of the image through convolution. In this embodiment, the feature maps are either stable or unstable and the tear film image V is fluorescent, ultraviolet, visible lighting, infrared or thermal. But the present invention is not limited to such application.
(18) Step 5 S5: classifying the eyelash images 21, the break-up area images 22, the non-break-up area images 23, the sclera images 24 and the eyelid images 25 output from the fully connected layer F through the softmax layer T. The eyelash images 21, the break-up area images 22, the non-break-up area images 23, the sclera images 24 and the eyelid images 25 respectively corresponds to a first prediction score P.sub.1, a second prediction score P.sub.2, a third prediction score P.sub.3, a fourth prediction score P.sub.4 and a fifth prediction score P.sub.5 to respectively produce a first label t.sub.1, a second label t.sub.2, a third label t.sub.3, a fourth label t.sub.4 and a fifth label t.sub.5.
(19) Step 6 S6: distinguishing the eyelash images 21, the break-up area images 22, the non-break-up area images 23, the sclera images 24 and the eyelid images 25 by the first label t.sub.1, the second label t.sub.2, the third label t.sub.3, the fourth label t.sub.4 and the fifth label t.sub.5 to detect a tear film break-up area 40 in the tear film image V as the image changes with time passing by and to quantize a tear film break-up time 50. In this embodiment, the tear film break-up time 50 is set at 5 seconds. Therefore, if a fluorescent tear film break-up time is equal to or less than 5 seconds, it is classified to a dry eye group, and if the break-up time is more than 5 seconds, it is classified to a normal group. But the present invention is not limited to such application.
(20) With reference to
(21) The tear film break-up time 50 is a stable indication of clinical diagnosis of eyes which represents the time period needed until the break-up appears after a blink. In this embodiment, the convolutional neural network model 10 is applied to detecting the tear film break-up time 50 after a training process of six layers of the model. Consequently, the present invention is able to detect and decide the stability of tear films objectively and conveniently as an auxiliary tool to ophthalmologists in clinical practices.
(22) Although particular embodiment of the invention has been described in detail for purposes of illustration, various modifications and enhancements may be made without departing from the spirit and scope of the invention. Accordingly, the invention is not to be limited except as by the appended claims.