Method and device for detecting an analyte in a body fluid
10309905 ยท 2019-06-04
Assignee
Inventors
Cpc classification
Y10T436/144444
GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
G01N2021/7769
PHYSICS
International classification
A61B1/00
HUMAN NECESSITIES
Abstract
A method for detecting at least one analyte in at least one sample of a body fluid is disclosed. Therein, at least one test element (124) is used, the at least one test element (124) having at least one test field (162) with at least one test chemistry (154) is used, wherein the test chemistry (154) is adapted to perform at least one optically detectable detection reaction in the presence of the analyte. The method comprises acquiring an image sequence of images of the test field (162) by using at least one image detector (178). Each image comprises a plurality of pixels. The method further comprises detecting at least one characteristic feature of the test field (162) in the images of the image sequence. The method further comprises correcting a relative position change between the image detector (178) and the test field (162) in the image sequence by using the characteristic feature, thereby obtaining a sequence of corrected images.
Claims
1. A method for detecting at least one analyte in at least one sample of a body fluid, wherein at least one test element with at least one test field is used, the at least one test field having at least one test chemistry, wherein the test chemistry is adapted to perform at least one optically detectable detection reaction in the presence of the analyte, wherein the method comprises acquiring an image sequence of images of the test field by using at least one image detector, wherein each image comprises a plurality of pixels, wherein the method further comprises detecting using a control unit at least one characteristic feature of the test field in the images of the image sequence, wherein the method further comprises correcting using the control unit a relative position change between the image detector and the test field in the image sequence by using the characteristic feature, thereby obtaining at least one corrected image.
2. The method according to claim 1, wherein the detecting of the characteristic feature comprises selecting at least one specific part of one or more images of the image sequence, denoting the information contained in this part as the characteristic feature, wherein other images of the image sequence are scanned or searched for this information or similar types of information.
3. The method according to claim 1, wherein the correction is individually adapted for each image of the image sequence, according to the characteristic feature detected in the specific image.
4. The method according to claim 1, wherein the correction of the relative position change comprises using at least one image of the image sequence as a reference image, wherein the reference image is kept unchanged, wherein the remaining images of the image sequence are corrected by using at least one calculational correction of the position of the pixels, wherein the calculational correction is chosen such that a correlation between the reference image and the corrected remaining images of the image sequence is maximized.
5. The method according to claim 4, wherein the calculational correction comprises at least one of the following: a shifting of the pixels of the remaining images of the image sequence in at least one spatial direction, wherein the shifting is chosen such that the correlation between the reference image and the corrected remaining images is maximized; or at least one rotation of the remaining images of the image sequence about at least one rotational axis by at least one rotation angle, wherein one or both of the rotational axis and the rotation angle are chosen such that the correlation between the reference image and the corrected remaining images is maximized.
6. The method according to claim 1, wherein the characteristic feature comprises at least one feature selected from the group consisting of: a roughness of the test field detectable in the images of the image sequence; a granularity of the test chemistry of the test field detectable in the images of the image sequence; faults of the test field detectable in the images of the image sequence; at least one fiducial mark comprised in the test field and detectable in the images of the image sequence.
7. The method according to claim 6, wherein the characteristic feature comprises at least two fiducial marks comprised in the test field and detectable in the images of the image sequence.
8. The method according to claim 1, wherein the sample of the body fluid is applied to the test field during acquisition of the image sequence, wherein at least one touchdown image is detected in the image sequence, wherein the touchdown image is an image of the image sequence acquired at a point in time closest to the moment of application of the sample of the body fluid onto the test field.
9. The method according to claim 1, wherein the sample of the body fluid is applied to the test field during acquisition of the image sequence, wherein the image sequence comprises a blank image sequence, wherein the blank image sequence comprises a plurality of blank images acquired before applying the sample of the body fluid to the test field, wherein at least one averaged blank image is derived from the blank images of the blank image sequence after performing the correction of the relative position change of the blank images of the blank image sequence.
10. The method according to claim 9, wherein the averaged blank image is derived in a continuous process during acquiring the images of the image sequence, wherein a preliminary averaged blank image is derived from the corrected blank images acquired so far, wherein new acquired blank images are used for revising the preliminary averaged blank image.
11. The method according to claim 1, wherein a moment of application of the sample of the body fluid onto the test field is detected in the image sequence.
12. The method according to claim 1, wherein after application of the sample of the body fluid onto the test field at least one region of interest is determined in the image sequence.
13. The method according to claim 12, wherein at least one corrected image acquired before or during application of the sample of the body fluid onto the test field is compared to at least one corrected image acquired after application of the sample of the body fluid onto the test field on a pixel-by-pixel basis, thereby generating a difference value for each pixel, wherein the difference value denotes a difference of the information contained in corresponding pixels of the corrected images acquired before or during and after application of the sample of the body fluid onto the test field, wherein the pixels are classified as pixels belonging to the region of interest or as pixels not belonging to the region of interest based on the difference values.
14. The method according to claim 12, wherein an image mask is generated denoting the pixels belonging to the region of interest.
15. A device for detecting at least one analyte in at least one sample of a body fluid, wherein the device comprises at least one test element receptacle for receiving at least one test element having at least one test field with at least one test chemistry, wherein the device further comprises at least one image detector for acquiring an image sequence of images of the test field, wherein the device further comprises at least one control unit, wherein the control unit is adapted to: acquire an image sequence of images of the test field by using at least one image detector, wherein each image comprises a plurality of pixels, detecting at least one characteristic feature of the test field in the images of the image sequence, and correct a relative position change between the image detector and the test field in the image sequence by using the characteristic feature, thereby obtaining at least one corrected image.
16. A test system for detecting at least one analyte in at least one sample of a body fluid, the test system comprising: at least one device for detecting at least one analyte in at least one sample of a body fluid and at least one test element having at least one test field with at least one test chemistry, wherein the test chemistry is adapted to perform at least one optically detectable detection reaction in the presence of the analyte, wherein the device comprises at least one test element receptacle for receiving at least one test element having at least one test field with at least one test chemistry, wherein the device further comprises at least one image detector for acquiring an image sequence of images of the test field, where-in the device further comprises at least one control unit, wherein the control unit is adapted to: acquire an image sequence of images of the test field by using at least one image detector, wherein each image comprises a plurality of pixels, detecting at least one characteristic feature of the test field in the images of the image sequence, and correct a relative position change between the image detector and the test field in the image sequence by using the characteristic feature, thereby obtaining at least one corrected image.
17. The test system according to claim 16, wherein the test system further comprises at least one puncture element, wherein the test system is adapted to puncture at least one skin portion of a user by using the puncture element, thereby creating the sample of the body fluid, wherein the test system is further adapted to transfer the sample of the body fluid onto the test field of the test element.
Description
BRIEF DESCRIPTION OF THE DRAWING FIGURES
(1) Further optional details and optional features of the present invention may be derived from the subsequent description of preferred embodiments, preferably in conjunction with the dependent claims. In these embodiments, in each case, the optional features may be realized in an isolated way or in an arbitrary combination of several features. The invention is not restricted to the embodiments. The embodiments are schematically depicted in the figures. Identical reference numbers in the figures refer to identical, similar or functionally identical elements.
(2) In the figures:
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19) Skilled artisans will appreciate that elements in the figures are illustrated for simplicity and clarity and have not necessarily been drawn to scale. For example, the dimensions of some of the elements in the figures may be exaggerated relative to other elements to help improve understanding of the embodiment(s) of the present invention.
DETAILED DESCRIPTION OF THE EMBODIMENTS
(20) For the purposes of describing and defining the present invention it is noted that terms like preferably, commonly, and typically are not utilized herein to limit the scope of the claimed invention or to imply that certain features are critical, essential, or even important to the structure or function of the claimed invention. Rather, these terms are merely intended to highlight alternative or additional features that may or may not be utilized in a particular embodiment of the present invention.
(21) In
(22) The device 112 may comprise one or more control units, which, in
(23) In the present embodiment, the magazine 114 comprises a plurality of test elements 124, received in the magazine 114 in a radial fashion, thereby providing an annular shape of the magazine 114 and/or a disk-shape of the magazine 114. It shall be noted, however, that other types of magazines 114 are possible and/or devices 112 using only one test element 124 rather than a plurality of test elements 124.
(24) The device 112 provides at least one application position 126. The device 112 is adapted to rotate the magazine 114 inside the receptacle 116 and to perform a test with the test element 124 located in the application position 126.
(25) Exemplary embodiments of the magazine 114 and/or the test elements 124 are disclosed in various views and details in
(26) Thus, the magazine 114 may comprise a magazine housing 128, which may also form part of housings 130 of the test elements 124. In this specific embodiment, the housing 130 comprises a lower shell 132, also referred to as the bottom part, which, typically, is made of an intransparent and preferably black plastics material. Further, the housing 130 comprises an upper shell 134, also referred to as the cover part, which, typically, is made of a transparent plastics material. Further, the housing 130 may comprise a sealing film 136, which typically is made of a metal foil, such as an aluminum foil, which may be glued to the upper shell 134 by an adhesive 138.
(27) Further, in this specific embodiment, each test element 124 may comprise one or more skin-piercing or puncture elements 140, which, as an example, may be formed as micro-samplers 142, each micro-sampler containing a lancet 144 with a lancet tip 146 and at least one capillary element, such as at least one capillary channel 148. Further potential details with regard to the micro-samplers 142 will be outlined below.
(28) Further, the magazine 114 may comprise a test chemistry ring 150 comprising a test chemistry carrier 152 and a test chemistry 154 applied to the test chemistry carrier 152 on a side facing the lower shell 132. The test chemistry ring 150 may be glued to the lower shell 132 by at least one adhesive 156, such as an adhesive tape, and/or maybe fixed to the magazine housing 128 by other means.
(29) Inside the magazine housing 128, a plurality of cavities 158 is formed, by appropriate recessions in the lower shell 132 and/or the upper shell 134. These cavities 158 may generally be oriented in a radial fashion, as depicted in
(30) In each cavity 158, further, a window 160 is formed in the lower shell 132. The test chemistry 154 accessible through these windows 160 thereby forms a test field 162 or part of a test field 162 for each test element 124. Thus, through the window 160, the sample of the body fluid may be applied to the test fields 162. Each test element 124 therefore, in the present embodiment, comprises at least one test field 162 and, optionally, a cavity 158, a puncture element 140 as well as a housing 130, which, in this embodiment, may be an integral part of the magazine housing 128.
(31) Further details of the sample generation and/or sample transfer will be explained with respect to
(32) As can be seen in
(33) In
(34) As an example, the light sources 182 may comprise one or more light-emitting diodes (LEDs), such as two light-emitting diodes, emitting in an ultraviolet or blue spectral range, such as in a spectral range of 350 to 400 nm, preferably in a spectral range of 350 to 380 nm or 360 to 365 nm. Alternatively or additionally, other commercially available LEDs, such as Green-LEDs (570+/30 nm); Red-LEDs (650+/50 nm) or IR-LEDs (700-1000 nm) may be employed. Additionally or alternatively to LEDs, one or more other types of light sources may be employed. Thus, as an example, light bulbs may be applied. Additionally or alternatively, typically depending on the requirements for the light signal, laser diodes may be used, even though this type of light sources typically implies increased costs.
(35) The detector 176 may further comprise one or more optical elements 184, such as one or more imaging optics, in order to image the test field 162 and/or at least one portion thereof onto the image detector 178, thereby creating an image 186 of the test field 162 and/or a part thereof on the image detector 178. The image 186 may comprise a matrix of information values, such as gray values, forming a matrix in one or two dimensions. In
(36) For the purpose of the sample transfer, as outlined above with regard to
(37) In
(38) As an example, CCD/CMOS image detectors 178 may be used, such as image sensors available from Eureca Messtechnik GmbH, Germany. Thus, image detectors of various manufacturers may be employed, such as CCD/CMOS image detectors manufactured by Fairchild imaging, Panavision, NEC, Sony, Toshiba, CMOS Sensor Inc., Kodak, Texas Instruments, TAOS or others. As an example, CCD/CMOS line sensors and/or areas sensors of one or more of models CCD111A, CCD424 manufactured by Fairchild imaging, of one or more of models LIS-500 or MDIC-2.0 manufactured by Panavision, of model PD3753CY-A manufactured by NEC, of one or more of models ICX207AK-E or ILX551B manufactured by Sony, of one or more types TCD1201DG or TCD132TG manufactured by Toshiba, of one or more of models M106-A9 or C106 manufactured by CMOS Sensor Inc., of one or more of models KAC9618 or KAC-01301 manufactured by Kodak, of model TC237B manufactured by Texas Instruments or of model TSL201R manufactured by TAOS may be used. Additionally or alternatively, camera boards containing one or more image sensor chips on printed circuit boards may be used as image detectors 178.
(39) As discussed in further detail above, the detector 176 may further comprise at least one wavelength-converting material, which is not depicted in the figures. Thus, the image detector 178 may be coated with one or more coatings comprising at least one wavelength-converting material such as at least one fluorescent material. Thus, specialized UV coatings having wavelength-converting properties are commercially available from Eureca Messtechnik GmbH, Germany. However, other types of wavelength-converting materials may be employed, such as fluorescent inorganic or organic materials.
(40) After wetting of the test field 162 by the sample of the body fluid, i.e. after application of the sample of the body fluid to the test field 162, the above-mentioned detection-reaction will take place, leading to optically detectable changes in the test field 162 and/or the test chemistry 154 contained therein. Examples of different images of the test field 162 as acquired by an image detector 178 are depicted in
(41) Thus, by evaluating the images 186, the concentration of the analyte may be determined, by directly or indirectly evaluating the information provided in a time-sequence of the images 186, which, herein, is referred to as an image sequence 186. Preferably, the image detector 178 may comprise a grid of photosensitive elements 180 having a dimension of 20 m to 50 m, preferably 30 m, in each direction. However, other dimensions are possible. Further, several photosensitive elements 180 of the image detector 178 may be combined to form combined photosensitive elements 180, wherein the information provided by these combined photosensitive elements 180 is combined and regarded as information of a superpixel of the image detector 178. In the present specification, this option shall be included, independent from the fact whether the raw photosensitive elements 180 of the image detector 178 are used or if several photosensitive elements 180 are combined, thereby creating an image detector comprising an array of superpixels.
(42) Typically, which is also possible within the present invention, only a portion of the images 186 is evaluated for determining the analyte concentration. Thus, a region of interest has to be defined, which defines the pixels of the image 186 which are considered for determining the analyte. In
(43) The option depicted in
(44) Therefore, as will be outlined in further detail below, a second option for determining the region of interest 190 is an analysis of the image sequence of the images 186 in an early phase of the wetting of the test field 162 with the sample of the body fluid and/or in an early phase of the process of the detection reaction. In this option, changes in the information contained in the pixels of the images 186 may be evaluated, which are caused by the wetting of the test field 162 after the transfer of the sample fluid. Specifically in case a signal-to-noise-ratio of the images 186 is sufficient, only wetted areas may be evaluated after the end point is reached, which may lead to a significant reduction of data storage volume and evaluation time.
(45) As a third option, which may be combined with the second option listed above, changes in the information values stored in the pixels of the images 186 of the image sequence may be evaluated for determining the region of interest. Thus, for detecting changes in the images 186, at least two of the images 186 may be compared, and the region of interest 190 may be determined on the basis of these detected changes. Thus, pixels of the images 186 may be selected based upon their history, such as by assigning those pixels with the highest rate of change in a certain time span to the region of interest 190. In
(46) As outlined above, the method according to the present invention comprises at least one correcting step correcting a relative position change between the image detector 178 and the test field 162 in the image sequence. As outlined above, the term relative position change may refer to any type of movement of the test field 162 as seen by the detector 176 and, specifically, by the image detector 178. This type of movement may be due to internal and/or external reasons in the test system 110. Thus, movements and corresponding position changes may be due to a handling of the test system 110, e.g. to mechanical vibrations during handling of the device 112 by a user, since, preferably, device 112 may be a hand-held device. Additionally or alternatively, movements may be due to the action of the test system 110 itself, i.e. to internal reasons. Thus, the application of the sample of the body fluid onto the test field 162, as depicted in
(47) According to the present invention, this relative position change between the image detector 178 and the test field 162 in the image sequence comprising images 186 acquired at different times is, at least partially, corrected. An example of a correction process will be explained with reference to
(48) Thus,
(49) As an example for a correction 198 based on the detection 196 of at least one characteristic feature, reference may be made to
(50) Thus, in
(51) Each image 186, including the reference image, may be described as a matrix comprising a number of information values I in each position or pixel of the image 186, such as as follows:
(52)
(53) Therein, I.sub.i,j denote the information values of the pixel i, j of the image I, such as gray values. N and M are integers denoting the width of the image 186 (N) and the height of the image (M). One specific position of this matrix, denoted by the coordinates i, j with 1iM and 1jN, denotes a specific pixel or position of the image 186.
(54) As indicated in
(55) For every possible value of the shift (r,s), a degree of conformity and/or a degree of identity or similarity is determined for the portion 200 and the corresponding portion of the image 186 to be searched. This is schematically depicted in
d.sub.E(r,s)=[.sub.(i,j)R(I(r+i,s+j)R(i,j).sup.2)].sup.1/2.
(56) By shifting the characteristic feature 202 (i.e. by shifting R) over the whole image 186 to be searched, one d.sub.E may be determined for each shift (r,s). Finally, by comparing all d.sub.E(r,s) determined this way, a minimum of all d.sub.E may be determined, i.e. a specific shift (r,s) may be determined for which d.sub.E assumes a minimum value. This shift denotes a best guess of a search result of the search for the characteristic feature 202 in the image 186. In order to avoid artifacts, this candidate of a shift may be compared to one or more limit values, i.e. by comparing the minimum value d.sub.E,min with at least one limit value. Only if d.sub.E,min is smaller or almost as big as the limit value, a positive match may be detected.
(57) It has to be noted, however, that the above-mentioned sum of squared differences is only one algorithm out of a large number of possible algorithms suited for searching for pattern matches for finding characteristic features in the image 186. This algorithm of finding pattern matches is e.g. disclosed in W. Burger et al.: Image Processing, Springer Verlag, London, 2008, pp. 429-436. However, additionally or alternatively, other types of pattern match algorithms searching for characteristic features in images 186 may be used, in order to determine a shift in between images.
(58) As soon as the search for the characteristic feature 202 in the image 186 was successful, the search will return a shift (r*,s*), indicating the amount of relative position change in between the image 186 and the reference image. This shift (r*,s*) may be used in method step 198 in
I*(i,j)=I(i+r*,j+s*), with 0i<M and 0j<N. For r*=0 and s*=0: I*=I.
(59) As an example, r* and s* may be limited to plausible values, such as values not exceeding 50. Instead of adding the shift (r*,s*), as indicated above, a subtraction is also possible.
(60) For further details of the potential algorithm for the correction step 194 and/or for further optional embodiments, reference may be made to the above-mentioned publication W. Burger et al.: Digital Image Processing, Springer Verlag, London, 2008, pp. 429-436. Specifically, the template matching algorithm disclosed in this text passage may be applied to the correction algorithm or the correction step 194. It should however be noted that other types of correlation and/or matching algorithms may be used, such as cross-correlation algorithms and/or pattern recognition algorithms. Further, it should be noted that the algorithm disclosed as an exemplary embodiment above, with regard to the examples provided in
(61) The whole correction step 194 in
(62) Further, as indicated by reference number 210 in
(63) The images 186 which are subject to the correction algorithm, such as the correction algorithm of
(64) Further, in conventional methods for qualitatively and/or quantitatively detecting an analyte concentration, the determination of a blank value and/or an empty value (both terms will be used as synonyms herein) typically plays an important role. Thus, since the optical properties of different patches of test fields 162 or test chemistries 154 may differ even in a dry state, the blank value may be used for normalizing detected optical changes which actually are due to the detection reaction. Typically, in known methods, such as in WO 2012/010454 A1, one or more blank values are acquired before applying the sample of the body fluid to the test field 162 and, after sample application, the subsequent measurement values are normalized by using this blank value, such as by dividing all subsequent measurement values rendered by the detector 176 by the at least one blank value.
(65) The present invention, specifically the correction step 194, offers the possibility of generating, at a very high precision, an averaged blank image rather than a single blank value, the averaged blank image containing averaged information of a plurality of blank images.
(66) In
(67) In a first step, a new image 186 is acquired by using the image detector 178, as denoted by method step 192 in
(68) Subsequently, in the newly acquired image or in the newly acquired, corrected image, at least one step 214 of detection of sample application is performed. This detection of sample application provides an answer to the question whether, in between the acquisition of the preceding image and the present, newly acquired image, the sample of the body fluid was applied to the test field 162. This step 214 of detection of sample application may be performed by detecting changes in the information values I(i,j) of the image or corrected image, as compared to the preceding image. As an example, changes of averages of the information values contained in the images or corrected images may be calculated and used, such as by using the following formula:
(69)
wherein |
(70) The difference averaged value |
(71)
(72) In the following, |I.sub.n,rel| is also referred to as
(73) Returning to the algorithm for detecting the averaged blank image in
(74) In case no sample application has been detected (branch N in
(75)
wherein B.sub.pr,n denotes the n.sup.th averaged blank image (pixel i,j), and I.sub.n denotes the newly acquired n.sup.th image before sample application (pixel i,j). As an initial value for B.sub.pr,1, the first blank image I.sub.1 may be used. Thus, an averaged blank image B.sub.pr,n may be generated by using a moving algorithm, by updating the preliminary averaged blank image B.sub.pr,n. Finally, as soon as the sample application has been detected (branch Y in
B(i,j)=B.sub.pr,n(i,j).
(76) This averaged blank image B may be used as a reference for all subsequent changes of the images which are due to the sample application.
(77) Thus, the averaged blank image B may be used for determining the analyte concentration by normalizing the images or corrected images, preferably after sample application, to the averaged blank image B on a pixel-by-pixel basis, such as by transforming the images (i.e. one image, a plurality of images or even all images) in one or both of the following transformed matrices:
I(i,j)=I(i,j)/B(i,j)
or
I(i,j)=I(i,j)B(i,j)
or
I(i,j)=(I(i,j)B(i,j))/B(i,j).
(78) Additionally or alternatively, as outlined above, at least one touchdown image T or corrected touchdown image may be used for determining the analyte concentration. Thus, as an example, one or more of the following transformed matrices may be used for determining the analyte concentration:
(79)
(80) The latter formula corresponds to the comparison matrix C.sub.n as defined above, which may also be used for detecting significant changes for the purpose of detecting a region of interest in the image sequence and/or the corrected image sequence.
(81) Other types of normalization processes are possible. In the following, when reference is made to the evaluation of the image sequence or corrected image sequence for the purpose of determining the analyte concentration, both the possibility of using the images or corrected images or the possibility of the normalized, transformed images or corrected images, such as by using one or more of the preceding formulae, shall be possible.
(82) Further, as outlined above, the determination of a region of interest plays an important role in many processes for detecting analytes in a body fluid. The method according to the present invention, specifically by creating the corrected image sequence, such as by using the algorithm depicted in
(83) Firstly, as depicted in
dI(i,j)=I.sub.m(i,j)I.sub.n(i,j),
wherein dI denotes a matrix indicating the change in information values and wherein I.sub.m denotes an image or corrected image or combined or transformed image acquired after the moment 218 of sample application and wherein I.sub.n denotes an image, a corrected image or a transformed or combined image acquired before or during the moment 218 of sample application. As an example, I.sub.n may be the above-mentioned touchdown image T. However, other embodiments are feasible, such as embodiments in which I.sub.n is an image acquired before the moment of sample application. Preferably, the images I.sub.m and I.sub.n are acquired as close as possible to the moment 218 of sample application. Thus, I.sub.n may be the image acquired immediately before the moment of sample application, and I.sub.m may be the image acquired immediately after sample application. Additionally or alternatively, images acquired at predetermined time distances before and after sample application may be compared, such as by using the image acquired one second before sample application as image I.sub.n and the image acquired one second after sample application as the image I.sub.m. Alternatively, I.sub.n may be the touchdown image, and I.sub.m may be an image acquired at a point in time 0.5 s to 4 s after the moment of sample application, such as 1 s after the moment of sample application. Further, several images may be combined, such as by using a preliminary averaged blank image instead of image I.sub.n and/or by using the averaged blank image B instead of image I.sub.n.
(84) In
(85) As can be seen in
(86) In order to define the region of interest 190 or a rough estimation of the region of interest 190, a threshold method may be used, for example by using an algorithm as depicted in
(87) Additionally or alternatively to the rough estimation of the region of interest by using the averaging threshold method depicted in
(88) Thus, the method depicted in
(89) Further, a histogram method may be used for evaluating the image of changes 246, as indicated by histogram 248 in
(90) Further, for evaluating the histogram 248, a further threshold method may be used. As outlined above, this threshold method may imply an automatic choice of one or more thresholds 250. For this purpose, threshold methods as known in the art may be used. Thus, preferably, the so-called Otsu method may be used. In this method, threshold 250 is chosen, separating the histogram 248 into two classes: class 252 of information values below threshold 250 and class 254 of information values above threshold 250, in the change matrix dI or a corrected change matrix dl, before or after filtering or applying additional data reduction steps. Threshold 250 may automatically be chosen such that the variance of values in each of the classes 252 is minimized, whereas the variance in between the values of different classes is maximized.
(91) In a next step, all pixels belonging to class 252 may be eliminated from the region of interest 190. Thus, a region of interest 190 in the form of a binary mask 256 may be generated, as depicted in the right part of
(92) In
(93) The region of interest 190 defined on a pixel-by-pixel basis, by using the binary mask 256, may be used for evaluating the images 186, preferably after correction step 194, such as by evaluating the corrected images acquired after the moment 218 of sample application. Thus, the corrected images 186 after performing the correction step 194 may be transformed as follows:
I.sub.ROI(i,j)=I(i,j).Math.ROI(i,j).
(94) Thereby, in any image, image sequence, group of images, corrected image or averaged image, all pixels outside the region of interest may be eliminated, whereas pixels inside the region of interest 190 may be kept unchanged. Thus, a masking of the images may take place.
(95) For further evaluation and determination of the analyte concentration, the pixels of the images inside the region of interest, such as the pixels of the matrix I.sub.ROI, may be evaluated. For this purpose, one or more of the images, corrected images or, for example, relative images such as one or more of the above mentioned images I, I or I may be masked, by using only the pixels of these images inside the region of interest. Thus, the above-mentioned image I may be used and masked for further evaluation, such as by using the following formula:
I.sub.ROI(i,j)=I(i,j).Math.ROI(i,j).
(96) Thereby, a matrix indicating a change in remission or percent relative remission may be created. From this matrix I.sub.ROI, average values over all pixels within the ROI may be created, wherein, basically, any type of averaging process may be used, such as median values over all pixels within the ROI, average values, weighted averages and other averaging processes.
(97) In
(98) The curves 260 to 268 as depicted in
(99) In
(100) Further, as outlined above with regard to
(101) Subsequently, in a series of further optional method steps, the reaction kinetics may be measured (step 284), the measurement results may be evaluated (step 286, analysis of measurement), and, further optionally, a statistical analysis of the measurement results may be performed (measurement statistics, step 288), before the method is ended (step 290).
(102) When looking at the method depicted in
(103) Having described the invention in detail and by reference to specific embodiments thereof, it will be apparent that modifications and variations are possible without departing from the scope of the invention defined in the appended claims. More specifically, although some aspects of the present invention are identified herein as preferred or particularly advantageous, it is contemplated that the present invention is not necessarily limited to these preferred aspects of the invention.
LIST OF REFERENCE NUMBERS
(104) 110 test system 112 device 114 magazine 116 receptacle 118 control unit 120 processor 122 user interface 124 test element 126 application position 128 magazine housing 130 housing 132 lower shell 134 upper shell 136 sealing film 138 adhesive 140 puncture element 142 micro-sampler 144 lancet 146 lancet tip 148 capillary channel 150 test chemistry ring 152 test chemistry carrier 154 test chemistry 156 adhesive 158 cavity 160 window 162 test field 164 engagement opening 166 actuator 168 puncture opening 170 viewing window 172 application side 174 detection side 176 detector 178 image detector 180 photosensitive element 182 light source 184 optical element 186 image 188 wetted portion 190 region of interest (ROI) 192 acquisition of new image 194 correction step 196 detect characteristic feature 198 correction 200 portion 202 characteristic feature 204 search region 206 create corrected image 208 repetition 210 further evaluation 212 boundaries 214 detection of sample application 216 peak 218 moment of sample application 220 no sample application detected 222 add new image to preliminary averaged blank image 224 sample application detected 226 define averaged blank image 228 background region 230 unwetted test field 232 region of significant changes 234 image of changes 236 average values of lines 238 average values of columns 242 threshold 244 threshold 246 image of changes 248 histogram 250 threshold 252 class of information values below threshold 254 class of information values above threshold 256 binary mask 258 bubbles or debris 260 20 mg/dl 262 70 mg/dl 264 150 mg/dl 266 250 mg/dl 268 550 mg/dl 270 start 272 detect test field 274 detect moment of sample application 276 detect blank image 278 detect significant changes 280 process significant changes 282 determine ROI 284 measure reaction kinetics 286 analysis of measurement 288 measurement statistics 290 end