METHOD FOR CHARACTERISING SPERM CELLS

20240044771 · 2024-02-08

Assignee

Inventors

Cpc classification

International classification

Abstract

A method for characterizing a mobile particle in a sample includes acquisition of at least one image of the sample during an acquisition period, using an image sensor and formation of a series of images, the series of images comprising at least one image; use of each image of the series of images as input image of a detection convolutional neural network, the detection convolutional neural network being configured to detect the particles and to produce, from each image, an output image on which each detected particle is assigned a distribution of intensity, centered on the particle and extending around the particle; for each detected particle, from each output image, estimation of a position of each detected particle in each image of the series of images; and characterization of each detected particle from the estimation of the position established from each image of the series of images.

Claims

1-18. (canceled)

19. A method for characterizing at least one mobile particle in a sample, the method comprising: a) acquisition of at least one image of the sample during an acquisition period, using an image sensor and formation of a series of images, the series of images comprising at least one image, each image of the sample being acquired according to a defocused imaging modality or according to a lensless imaging modality, such that each particle forms, on each image, a diffraction figure; b) use of each image of the series of images resulting from a) as input image of a detection convolutional neural network, the detection convolutional neural network being configured to detect the particles and to produce, from each image, an output image on which each detected particle is assigned a distribution of intensity, centered on the particle and extending around the particle; c) for each detected particle, from each output image resulting from b), estimation of a position of each detected particle in each image of the series of images; d) characterization of each detected particle from the estimation of the position resulting from c), established from each image of the series of images.

20. The method as claimed in claim 19, wherein the particle is a sperm cell.

21. The method as claimed in claim 20, wherein the step d) comprises a morphological characterization of each detected sperm cell, the step d) comprising: for each detected sperm cell, from each image resulting from a), and from the positions resulting from c), extraction of a thumbnail image comprising the detected sperm cell, the position of the detected sperm cell in the thumbnail image being predetermined, so as to obtain, for a detected sperm cell, a series of thumbnail images, the size of each thumbnail image being less than the size of each image acquired in the step a); for each detected sperm cell, use of the series of thumbnail images as input data of a classification neural network, the classification neural network being configured to classify the sperm cell from among predetermined morphological classes.

22. The method as claimed in claim 21, wherein each detected sperm cell is centered with respect to each thumbnail image.

23. The method as claimed in claim 21, wherein the step d) comprises a characterization of the motility of the sperm cell, the step d) comprising, from positions of the sperm cell resulting from the step c): a determination of a trajectory of the sperm cell during the acquisition period; a calculation of a speed of the sperm cell from the trajectory; a classification of the sperm cell as a function of the speed.

24. The method as claimed in claim 23, comprising: a determination of a distance, in a straight line, between a first point and a last point of the trajectory, and a calculation of a linear trajectory speed from said distance; and/or a determination of distances traveled between each acquired image, and a calculation of a curvilinear trajectory speed from said distances; and/or a determination of a smoothed trajectory, and a calculation of an average trajectory speed from the smoothed trajectory.

25. The method as claimed in claim 19, wherein: the sample extends on a plane of the sample; the image sensor extends on a detection plane; an optical system extends between the sample and the image sensor, the optical system defining an object plane and an image plane; the object plane is offset with respect to the plane of the sample by an object defocusing distance and/or the image plane is offset with respect to the plane of the sample by an image defocusing distance, such that, in the step a), each image of the sample is acquired according to a defocused imaging modality.

26. The method as claimed in claim 25, wherein: the sample is disposed on a sample support, resting on at least one spring, the spring being configured to push the sample support toward the optical system; the optical system is linked to at least one stop, extending, from the optical system, toward the sample support; such that, during the step a), under the effect of the spring, the sample support is pressed onto the stop.

27. The method as claimed in claim 19, wherein no image-forming optic extends between the sample and the image sensor, such that, in the step a), each image of the sample is acquired according to a lensless imaging modality.

28. The method as claimed in claim 19, wherein the step a) comprises a normalization of each acquired image by an average of said image or by an average of images of the series of images.

29. The method as claimed in claim 19, wherein the step a) comprises an application of a high-pass filter to each acquired image.

30. The method as claimed in claim 19, wherein, in the step b), the distribution of intensity assigned to each particle is decreasing, such that the intensity decreases as a function of the distance with respect to the particle.

31. The method as claimed in claim 19, wherein, in the step b), the distribution of intensity assigned to each particle is a two-dimensional parametric statistical distribution.

32. A device for observing a sample, the sample comprising mobile particles, the device comprising: a light source, configured to illuminate the sample; an image sensor, configured to form an image of the sample; a holding structure, configured to hold the sample between the light source and the image sensor; a processing unit, linked to the image sensor, and configured to implement the steps a) to d) of a method as claimed in claim 19 from at least one image acquired by the image sensor.

33. The device as claimed in claim 32, wherein no image-forming optic extends between the image sensor and the sample.

34. The device as claimed in claim 33, wherein the holding structure is configured to maintain a fixed distance between the image sensor and the sample.

35. The device as claimed in claim 32, wherein: the image sensor extends on a detection plane; the device comprises an optical system, extending between the image sensor and the support plane, the optical system defining an image plane and an object plane, the device being such that: the image plane is offset with respect to the detection plane by an image defocusing distance; and/or the support plane is offset with respect to the object plane by an object defocusing distance.

36. The device as claimed in claim 35, wherein: the holding structure comprises a sample support, mounted on a spring; the optical system is linked mechanically to at least one stop; the spring is arranged such that the sample support is pressed against the stop, so as to maintain a distance between the optical system and the sample.

Description

FIGURES

[0060] FIG. 1A represents a first embodiment of a device allowing an implementation of the invention.

[0061] FIG. 1B is a three-dimensional view of the device schematically represented in FIG. 1A.

[0062] FIG. 1C shows a setup making it possible to maintain the sample at a fixed distance from an optical system.

[0063] FIG. 1D shows the setup represented in FIG. 1C in a position of observation.

[0064] FIG. 2 represents a second embodiment of a device allowing an implementation of the invention.

[0065] FIG. 3 shows the main steps of a method for characterizing mobile particles in the sample.

[0066] FIGS. 4A, 4B and 4C respectively represent an acquired image, a reference image and an image resulting from the detection neural network.

[0067] FIG. 5A is an example of image of a sample obtained by implementing a device according to the first embodiment of the invention. FIG. 5B shows the image represented in FIG. 5A after processing by a detection neural network. FIG. 5C shows an example of obtaining trajectories of particles detected in the image 5B.

[0068] FIG. 6 shows thumbnail images centered on a same sperm cell, the thumbnail images being extracted from a series of nine images acquired according to a defocused imaging modality.

[0069] FIGS. 7A, 7B and 7C respectively represent an acquired image, a reference image and an image resulting from the detection neural network.

[0070] FIGS. 8A, 8B and 8C respectively represent an acquired image, a reference image and an image resulting from the detection neural network. On this series of images, FIG. 8A has been subjected to processing by a high-pass filter.

[0071] FIGS. 9A and 9B represent particle detection performance levels by the detection neural network, in different conditions. The performance levels are the sensitivity (FIG. 9A) and the specificity (FIG. 9B).

[0072] FIGS. 10A to 10C represent confusion matrices, expressing the sperm cell classification performance levels, using images acquired according to the defocused modality, the classification being performed respectively: [0073] by implementing a holographic reconstruction algorithm from an image obtained by holographic reconstruction and processed by a neural network; [0074] by implementing a classification neural network at which the input layer comprises a single image acquired according to a defocused configuration, without holographic reconstruction;

[0075] by implementing a classification neural network, of which the input layer comprises a series of images of a particle, each image being acquired according to a defocused configuration, without holographic reconstruction.

EXPLANATION OF PARTICULAR EMBODIMENTS

[0076] FIG. 1A shows a first embodiment of a device allowing an implementation of the invention. According to this first embodiment, the device allows the observation of a sample 10 interposed between a light source 11 and an image sensor 20. The light source 11 is configured to emit an incident light wave 12 propagated to the sample parallel to a propagation axis Z.

[0077] The device comprises a sample support 10s configured to receive the sample 10, such that the sample is held on the support 10s. The sample thus held extends on a plane, called sample plane P.sub.10. The sample plane corresponds for example to an average plane around which the sample 10 extends. The sample support can be a glass plate, for example 1 mm thick.

[0078] The sample notably comprises a liquid medium 10.sub.m in which mobile and possibly motile particles 10.sub.i are bathed. The medium 10.sub.m can be a biological liquid or a buffer liquid. It can for example comprise a bodily fluid, in the pure or diluted state. Bodily fluid should be understood to mean a liquid generated by a living body. It can in particular be, as a nonlimiting example, blood, urine, cerebrospinal fluid, semen, lymph.

[0079] The sample 10 is preferably contained in a fluidic chamber 10.sub.c. The fluidic chamber is for example a fluidic chamber of a thickness of between 20 m and 100 m. The thickness of the fluidic chamber, and therefore of the sample 10, on the propagation axis Z, varies typically between 10 m and 200 m, and preferably lies between 20 m and 50 m.

[0080] One of the objectives of the invention is the characterization of particles in motion in the sample. In the exemplary embodiment described, the mobile particles are sperm cells. In this case, the sample comprises semen, possibly diluted. In this case, the fluidic chamber 10.sub.c can be a counting chamber dedicated to analyzing the mobility or the concentration of cells. It may for example be a counting chamber marketed by Leja, of a thickness of between 20 m and 100 m.

[0081] According to other applications, the sample comprises mobile particles, for example microorganisms, for example microalgae or plankton, or cells, for example cells in the process of sedimentation.

[0082] The distance D between the light source 11 and the sample 10 is preferably greater than 1 cm. It is preferably between 2 and 30 cm. Advantageously, the light source 11, seen by the sample, is considered to be a spot source. That means that its diameter (or its diagonal) is preferentially less than a tenth, better a hundredth, of the distance between the sample and the light source.

[0083] The light source 11 is for example a light-emitting diode. It is preferably

[0084] associated with a diaphragm 14, or spatial filter. The aperture of the diaphragm is typically between 5 m and 1 mm, preferably between 50 m and 1 mm. In this example, the diaphragm has a diameter of 400 m. According to another configuration, the diaphragm can be replaced by an optical fiber, a first end of which is placed facing the light source and a second end of which is placed opposite the sample 10. The device can also comprise a diffuser 13, disposed between the light source 13 and the diaphragm 14. The use of a diffuser/diaphragm assembly is for example described in U.S. Pat. No. 10,418,399.

[0085] The image sensor 20 is configured to form an image of the sample on a detection plane P.sub.20. In the example represented, the image sensor 20 comprises a matrix of pixels, of CCD or CMOS type. The detection plane P.sub.20 preferably extends at right angles to the propagation axis Z. Preferably, the image sensor has a high sensitive surface area, typically greater than 10 mm.sup.2. In this example, the image sensor is an IDS-UI-3160CP-M-GL sensor comprising pixels of 4.84.8 m.sup.2, the sensitive surface area being 9.2 mm5.76 mm, i.e. 53 mm.sup.2.

[0086] In the example represented in FIG. 1A, the image sensor 20 is optically coupled to the sample 10 by an optical system 15. In the example represented, the optical system comprises a lens 15.sub.1 and a tube lens 15.sub.2. The latter is intended to project a formed image onto the sensitive surface of the image sensor 20 (surface area of 53 mm.sup.2).

[0087] In this example: [0088] the lens 15.sub.1 is a Motic CCIS EF-N Plan Achromat 10 lens, with a numerical aperture of 0.25; [0089] the lens 15.sub.2 is a Thorlabs LBF254-075-A lens with a focal length of 75 mm.

[0090] Such a setup gives an observation field of 3 mm.sup.2, with a spatial resolution of 1 m. The short focal length of the lens makes it possible to optimize the bulk of the device 1 and to adjust the magnification to the dimension of the image sensor.

[0091] The image sensor is configured to acquire images of the sample, according to an acquisition frequency of a few tens of images per second, for example 60 images per second. The sampling frequency is typically between 5 and 100 images per second.

[0092] The optical system 15 defines an object plane P.sub.o and an image plane P.sub.i. In the embodiment represented in FIG. 1A, the image sensor 20 is configured to acquire an image according to a defocused configuration. The image plane P.sub.i is merged with the detection plane P.sub.20, while the object plane P.sub.o is offset by an object focusing distance of between 10 m and 500 m, with respect to the sample. The focusing distance is preferably between 50 m and 100 m, for example 70 m. The object plane P.sub.o extends out of the sample 10. According to another possibility, the object plane extends into the sample, while the image plane is offset with respect to the detection plane, by an image defocusing distance. The image focusing distance is preferably between 50 m and 100 m, for example 70 m. According to another possibility, the object plane P.sub.o and the image plane P.sub.i are both offset respectively with respect to the plane of the sample and with respect to the detection plane. Whatever the configuration retained, the defocusing distance is preferably greater than 10 m and less than 1 mm, even 500 m, and preferably between 50 m and 150 m. The observation of a cellular sample according to a defocused configuration has been described in the patent U.S. Pat. No. 10,545,329.

[0093] According to such a modality, the image sensor 20 is exposed to a light wave, called exposure light wave. The image acquired by the image sensor comprises interference figures, that can also be referred to by the term diffraction figures, formed by: [0094] a portion of the light wave 12 emitted by the light source 11, and having passed through the sample without interacting therewith; [0095] diffraction waves, formed by the diffraction of a portion of the light wave 12 emitted by the light source in the sample. The diffraction is notably that formed by the particles.

[0096] A processing unit 30, comprising a microprocessor for example, is able to process each image acquired by the image sensor 20. In particular, the processing unit comprises a programmable memory 31 in which is stored a sequence of instructions for performing the image processing and computation operations described in this description. The processing unit 30 can be coupled to a screen 32 allowing the display of images acquired by the image sensor 20 or resulting from the processing performed by the processing unit 30.

[0097] The image acquired by the image sensor 20, according to a defocused imaging modality, is a diffraction figure of the sample, sometimes called hologram. It does not make it possible to obtain an accurate representation of the observed sample. Usually in the field of holography, it is possible to apply, to each image acquired by the image sensor, a holographic reconstruction operator so as to calculate a complex expression representative of the light wave to which the image sensor is exposed, and that can be done at any point of coordinates (x, y, z) of the space, and in particular in a reconstruction plane corresponding to the plane of the sample. The complex expression makes it possible to obtain the intensity or the phase of the exposure light wave. Such a holographic reconstruction is described in association with the prior art, and in U.S. Pat. No. 10,545,329.

[0098] However, for processing speed reasons, the inventors have followed an approach that differs from that suggested by the prior art, without recourse to a holographic reconstruction algorithm applied to the images acquired by the image sensor. The method implemented by the sensor is described hereinbelow, in association with FIG. 3.

[0099] FIG. 1B is a representation of an example of device as schematically represented in FIG. 1A. FIGS. 1C and 1D represent a detail of the arrangement of the sample 10 facing the optical system 15, and more specifically facing the lens 15.sub.1. The support of the sample 10.sub.s is linked to elastic return means 16.sub.a, for example springs. The springs tend to bring the sample support 10.sub.s closer to the lens 15.sub.1. The device also comprises stops 16.sub.b, secured to the lens 15.sub.1. The lens 15.sub.1 and the stops 16.sub.b are fixed with respect to the image sensor 20. The stops 16.sub.b and the return means 16.sub.a form a holding structure 16, intended to hold the sample between the light source 11 and the image sensor 20.

[0100] FIGS. 1C and 1D represent the holding of the sample on the sample support respectively during the placement of the sample and during the observation thereof. A rigid link 16.sub.c links the lens 15.sub.1 to the stop 16.sub.b. During the placement of the sample (FIG. 1C), the springs 16.sub.a are compressed. During the observation of the sample (FIG. 1D), under the effect of an expansion of the springs 16.sub.a, the support of the sample 10.sub.s comes to bear against the stops 16.sub.b. That makes it possible to control the distance between the lens 15.sub.1 and the sample 10, independently of the thickness of the sample support 10.sub.s. That makes it possible to overcome variations of the thickness of the sample support. For example, when the latter is a glass plate 1 mm thick, the thickness can fluctuate by a relatively significant range, for example 100 m. The setup described in association with FIGS. 1C and 1D makes it possible to control the distance between the lens 15.sub.1 and the sample 10, to within 5 m, independently of such fluctuations of thickness of the plate 10.sub.s. The inventors estimate that such a setup makes it possible to avoid recourse to an autofocus system for taking images.

[0101] In FIGS. 1C and 1D, also represented is a heating resistor 19 linked to a temperature controller 18. The function of these elements is to maintain a temperature of the fluidic chamber 10.sub.c at 37 C.

[0102] FIG. 2 represents a second embodiment of the device 1 suitable for the implementation of the invention. The device 1 comprises a light source 11, a diffuser 13, a diaphragm 14, an image sensor 20, a holding structure 17 and a processing unit 30 as described in association with the first embodiment. The holding structure 17 is configured to define a fixed distance between the sample and the image sensor. According to this embodiment, the device does not include an image-forming lens between the image sensor 20 and the sample 10. The image sensor 20 is preferably brought close to the sample, the distance between the image sensor 20 and the sample 10 being typically between 100 m and 3 mm. According to this embodiment, the image sensor acquires images according to a lensless imaging modality. The sample is preferably contained in a fluidic chamber 10.sub.c, for example a Leja chamber as described in association with the first embodiment. The advantage of such an embodiment is that it does not require a precise positioning of an optical system 15 with respect to the sample 10, and that it gives a high observation field. The drawback is that the images obtained are of lesser quality, but remain usable.

[0103] Preferably, the holding structure is arranged such that the distance between the sample, when the sample 10 is disposed on the holding structure 17, and the image sensor 20, is constant. In the example represented in FIG. 2, the holding structure 17 is linked to a base 20, on which the image sensor 20 extends. The base can for example be a plate of PCB (printed circuit board) type, on which the image sensor 20 is placed. The sample 10, contained in the fluidic chamber 10.sub.c, is held, by the sample support 10.sub.s, on the holding structure 17. The sample support is, for example, a transparent plate. The sample extends between the sample support 10.sub.s and the image sensor. Thus, the distance between the sample 10 and the image sensor 20 is not affected by a fluctuation of the thickness of the support 10s.

[0104] FIG. 3 schematically represents the main steps of a method for processing several images acquired by an image sensor according to the defocused imaging modality (first embodiment) or lensless imaging modality (second embodiment). The method is described in association with the observation of sperm cells, it being understood that it can be applied to the observation of other types of motile particles.

[0105] Step 100: Acquisition of a series of images I.sub.0,n

[0106] During this step, a series of images I.sub.0,n is acquired according to one of the modalities previously described. n is a natural integer designating the rank of each image acquired, with 1nN, N being the total number of images acquired. The images acquired are images acquired either according to a defocused imaging modality or according to a lensless imaging modality. The number N of images acquired can be between 5 and 50. As previously indicated, the images can be acquired according to an acquisition frequency of 60 Hz.

[0107] In the results presented hereinbelow, the images were acquired by using a defocused imaging modality, as described in association with FIGS. 1A to 1D.

[0108] Step 110: Preprocessing

[0109] The aim is to perform a preprocessing of each image acquired, so as to limit the effects of a fluctuation of the intensity of the incident light wave or of the sensitivity of the camera.

[0110] The preprocessing consists in normalizing each image acquired by an average of the intensity of at least one image acquired, and preferably all of the images acquired.

[0111] Thus, from each image I.sub.0,n, the normalization makes it possible to obtain a normalized image I.sub.n, such that

[00001] I n = I 0 , n I 0 , n _ or I n = I 0 , n I 0 , n _ 100

in which I.sub.0,n is an average of one or more images of the series of images, and preferably of each image of the series of images.

[0112] According to one possibility, the preprocessing can include an application of a high-pass filter to each image, possibly normalized. The high-pass filter makes it possible to eliminate the low frequencies of the image.

[0113] According to one possibility, a Gaussian filter is applied, by effecting a product of convolution of the image I.sub.n by a Gaussian kernel K. The width at mid-height of the Gaussian kernel is for example 20 pixels. The preprocessing then consists in subtracting, from the image I.sub.n, the image I.sub.n*K resulting from the application of the Gaussian filter, such that the image resulting from the preprocessing is: I.sub.n=I.sub.nI.sub.n* K. * is the convolution product operator.

[0114] The effect of such a filtering is described hereinbelow, in association with FIGS. 7A to 7C and 8A to 8C.

[0115] Step 120: Detection of Particles

[0116] This step consists in detecting and in accurately positioning the sperm cells, and doing so on each image I.sub.0,n, resulting from the step 100 or of each image I.sub.n preprocessed in the step 110. This step is performed using a detection convolutional neural network CNN.sub.d. The neural network CNN.sub.d comprises an input layer, in the form of an input image I.sub.in,n. The input image I.sub.in,n is either an acquired image I.sub.0,n, or a preprocessed image I.sub.n, I.sub.n. From the input image I.sub.in,n, the detection neural network CNN.sub.d generates an output image I.sub.out,n. The output image I.sub.out,n is such that each sperm cell 10.sub.i detected on the input image I.sub.in,n, in a position (x.sub.i, y.sub.i), appears in the form of a distribution of intensity centered around said position. In other words, from an input image I.sub.in,n, the neural network CNN.sub.d allows: [0117] a detection of sperm cells; [0118] an estimation of the position (x.sub.i, y.sub.i) of each detected sperm cell 10.sub.i; [0119] a generation of an output image I.sub.out,n comprising a distribution of intensity D.sub.i predetermined around each position (x.sub.i, y.sub.i).

[0120] Thus, I.sub.out,n=CNN.sub.d(I.sub.in,n)

[0121] Each position (x.sub.i, y.sub.i) is a two-dimensional position, in the detection plane P.sub.20. The distribution of intensity D.sub.i is such that the intensity is maximal at each position (x.sub.i, y.sub.i) and that the intensity is considered negligible beyond a vicinity V.sub.i of each position. Vicinity V.sub.i is understood to mean a region extending according to a number of pixels, predetermined, for example between 5 and 20 pixels, around each position (x.sub.i, y.sub.i). The distribution D.sub.i can be of strobe type in which case, in the vicinity V.sub.i, each pixel is of constant intensity is high, and beyond the vicinity V.sub.i, each pixel has a zero intensity.

[0122] Preferably, the distribution D.sub.i is centered on each position (x.sub.i, y.sub.i) and is strictly

[0123] decreasing around the latter. It can for example be a two-dimensional Gaussian intensity distribution, centered on each position (x.sub.i, y.sub.i). The width at mid-height is for example less than 20 pixels, and preferably less than 10 or 5 pixels. Any other form of parametric distribution can be envisaged, bearing in mind that it is preferable for the distribution D.sub.i to be symmetrical around each position, and preferably strictly decreasing from the position (x.sub.i, y.sub.i). The fact of assigning an intensity distribution D.sub.i to each position (x.sub.i, y.sub.i) makes it possible to obtain an output image I.sub.out,n in which each sperm cell is simple to detect. In that, the output image I.sub.out,n is a detection image, on the basis of which each sperm cell can be detected.

[0124] The output image from the neural network is formed by a resultant of each intensity distribution D.sub.i defined respectively around each position (x.sub.i, y.sub.i).

[0125] In the example implemented by the inventors, the detection convolutional neural network comprised 20 layers comprising either 10 or 32 characteristics (more usually referred to as features) per layer. The number of features was determined empirically. The transition from one layer to another is performed by applying a convolution kernel of 33 size. The output image is obtained by combining the features of the last convolution layer. The neural network was programmed in a Matlab environment (Matlab publisher: The Mathworks). The neural network was previously subjected to learning (step 80), so as to parameterize the convolution filters. During the learning, learning sets were used, each set comprising: [0126] an input image, obtained by the image sensor and preprocessed by normalization and possibly filtering, by using animal semen, [0127] an output image, on which the position of each sperm cell was annotated manually.

[0128] The learning was performed by using 10 000 or 20 000 annotated positions (i.e. between 1000 and 3000 annotated positions per image). The effect of the size of the learning set (10 000 or 20 000 annotations) is discussed in association with FIGS. 9A and 9B.

[0129] FIGS. 4A, 4B and 4C are a set of test images, respectively comprising: [0130] an image of a sample of bovine semen acquired by the image sensor and having been subjected to a normalization and high-pass filtering according to the step 110; [0131] a manually annotated image, which corresponds to a reference image, usually designated by the term ground truth; [0132] an image resulting from the application of the detection convolutional neural network.

[0133] The image of FIG. 4C is consistent with the reference image (FIG. 4B) which affirms the detection performance of the neural network CNN.sub.d. The inventors have noted that FIG. 4C shows a position not represented on the reference image: a check showed that it was due to an annotation being forgotten on the reference image.

[0134] Thus, on the basis of a hologram, and through possible preprocessings of normalization or high-pass filter type, to reveal the high frequencies, the application of the detection convolutional neural network CNN.sub.d allows a detection and an accurate positioning of sperm cells. And does so without recourse to an image reconstruction implementing a holographic propagation operator, as suggested in the prior art. The input image of the neural network is an image formed in the detection plane, and not in a reconstruction plane remote from the detection plane. Furthermore, the detection neural network generates output images with little noise: the signal-to-noise ratio associated with each sperm cell detection is high, which facilitates the subsequent operations.

[0135] The step 120 is reiterated for different images of a same series of images, so as to obtain an output series of images I.sub.out,1 . . . I.sub.out,N.

[0136] During the step 120, from each output image I.sub.out,1 . . . I.sub.out,N, a local maximum detection algorithm is applied, so as to obtain, for each image, a list of 2D coordinates, each coordinate corresponding to a position of a sperm cell. Thus, from each output image I.sub.out,1 . . . I.sub.out,N, a list L.sub.out,1 . . . L.sub.out,N is established. Each list L.sub.out,n comprises corresponds to the positions 2D, in the detection plane, of sperm cells detected in an image I.sub.out out,n.

[0137] FIG. 5A represents an image acquired in the step 100. More specifically, it is a detail of an image acquired according to a defocused modality. It is an image of a sample comprising bovine sperm. Such an image is difficult to interpret by a user. FIG. 5B shows an image resulting from the application of the detection neural network to FIG. 5A, after normalization and application of a high-pass filter. The comparison between the images 5A and 5B shows the gain provided by the detection convolutional neural network in terms of signal-to-noise ratio.

Step 130: Position Tracking

[0138] During this step, from the lists L.sub.out,1 . . . L.sub.out,N respectively established from the output images I.sub.out,1 . . . I.sub.out,N, resulting from a same series of images I.sub.0,1 . . . I.sub.0,N, a position tracking algorithm is applied, normally referred to as tracking algorithm. For each sperm cell detected following the step 120, an algorithm is implemented that allows the position of the sperm cell to be tracked, parallel to the detection plane, between the different images of the series of images. The implementation of the position tracking algorithm is efficient because it is performed from the lists L.sub.out,1 . . . L.sub.out,N resulting from the step 120. As previously indicated, these images have a high signal-to-noise ratio, which facilitates the implementation of the position tracking algorithm. The position tracking algorithm can be an algorithm of closest neighbor type. The step 130 allows a determination of a trajectory of each sperm cell parallel to the detection plane.

[0139] In FIG. 5C, output images I.sub.out,1 . . . I.sub.out,N have been superposed by considering a stack of images of 30 images. It can be seen that the trajectory, parallel to the detection plane, of each sperm cell, can be determined precisely.

[0140] Step 140: Characterization of Motility

[0141] During the step 140, for each sperm cell, each trajectory can be characterized on the basis of metrics applied to the trajectories resulting from the step 130. Knowing the image acquisition frequency, it is possible to quantify the speeds of displacement of each sperm cell detected, and in particular: [0142] a linear trajectory speed VSL, usually referred to as velocity straightline path, which corresponds to the speed calculated on the basis of a distance, in a straight line, between the first and last points of the trajectory (the first point is determined from the first acquired image of the series of images, and the last point is determined from the last acquired image of the series of images), [0143] a curvilinear trajectory speed VCL, usually referred to as velocity curvilinear path: this is a speed established by aggregating the distances traveled between each image, and by dividing by the duration of the acquisition period; [0144] an average trajectory speed VAP, usually referred to as velocity average path: this is a speed established after smoothing of the trajectory of a particle: the distance traveled according to the smoothed trajectory (or average trajectory) is divided by the duration of the acquisition period.

[0145] From the calculated speeds, it is possible to define indicators making it possible to characterize the motility of the sperm cells that are known to the person skilled in the art. These are for example indicators of the following types: [0146] straightness indicator STR, obtained by a ratio between VSL and VAP, usually referred to as straightness. This indicator is all the higher when the sperm cell moves in a straight line; [0147] linearity indicator LIN, obtained by a ratio between VSL and VCL, usually referred to as linearity. This indicator is also all the higher when the sperm cell moves in a straight line; [0148] wobble indicator WOB, obtained by a VAP/VCL ratio, usually referred to as wobble.

[0149] The quantification of the speeds or parameters listed above makes it possible to categorize the sperm cells according to their motility. For example, a sperm cell is considered as: [0150] motile if the length of the trajectory is greater than a first threshold, for example 10 pixels, and its movement along the average trajectory (VAPxt, t being the acquisition period) is greater than a predefined length, for example corresponding to the length of a sperm cell head; [0151] progressive if the length of the trajectory is greater than the first threshold, and if the straightness STR and the average trajectory speed VAP are respectively greater than two threshold values STR.sub.th and VAP.sub.th1, [0152] slow if the length of the trajectory is greater than the first threshold and if the linear trajectory speed VSL and the average trajectory speed VAP are respectively less than two threshold values VSL.sub.th2 and VAP.sub.th2; [0153] static if the length of the trajectory is greater than the first threshold and if the linear trajectory speed VSL and the average trajectory speed VAP are respectively less than two threshold values VSL.sub.th3 and VAP.sub.th3; [0154] not categorized if the length of the trajectory is less than the first threshold.

[0155] The threshold values STR.sub.th, VSL.sub.th2, VSL.sub.th3 are previously determined, with VSL.sub.th2VSL.sub.th3. The same applies for the values VAP.sub.th1, VAP.sub.th2 and VAP.sub.th3 with VAP.sub.th1VAP.sub.th1VAP.sub.th3.

[0156] Step 150: Extraction of Thumbnail Images for Each Sperm Cell

[0157] During this image, for each sperm cell detected by the neural network CNN.sub.d, a thumbnail image V.sub.i,n extracted, and from each image I.sub.0,n acquired by the image sensor in the step 100.

[0158] Each thumbnail image V.sub.i,n is a portion of an image I.sub.0,n acquired by the image sensor. For each detected sperm cell 10.sub.i, from each image I.sub.0,n, a thumbnail image V.sub.i,n is extracted around the position (x.sub.i, y.sub.i) assigned to the sperm cell. The position (x.sub.i, y.sub.i) of the sperm cell 10.sub.i, in each image I.sub.0,n, is obtained following the implementation of the position tracking algorithm (step 130).

[0159] The size of each thumbnail image V.sub.i,n is predetermined. Relative to each thumbnail image V.sub.i,n, the position (x.sub.i, y.sub.i) of the sperm cell 10.sub.i considered is fixed: preferably, the position (x.sub.i, y.sub.i) of the sperm cell 10.sub.i is centered in each thumbnail image V.sub.i,n.

[0160] A thumbnail image V.sub.i,n can for example comprise several tens or even hundreds of pixels, typically between 50 and 500 pixels. In this example, a thumbnail image comprises 6464 pixels. Because of the size, and of the concentration of the sperm cells in the sample, a thumbnail image V.sub.i,n can include several sperm cells to be characterized. However, only the sperm cell 10.sub.i occupying a predetermined position in the thumbnail image V.sub.i,n, for example at the center of the thumbnail image, is characterized by using said thumbnail image.

[0161] FIG. 6 shows nine thumbnail images V.sub.i,n extracted from images I.sub.out,n resulting from the step 100, by using the positions (x.sub.i, y.sub.i) resulting from the sperm cell trajectory tracking resulting from the step 130.

[0162] Step 160: Classification of the Morphology of Each Sperm Cell

[0163] During this step, a classification neural network CNN.sub.c is used, so as to classify each sperm cell 10.sub.i previously detected by the detection convolutional neural network CNN.sub.d. For each sperm cell 10.sub.i, the neural network is fed by the thumbnail images V.sub.i,1 . . . V.sub.i,N extracted in the step 150.

[0164] The classification convolutional neural network CNN.sub.c can for example comprise 6 convolutional layers, comprising between 16 and 64 features: 16 features for the four first layers, 32 features for the fifth layer, and 64 features for the sixth layer. The transition from one layer to another is performed by applying a convolution kernel of 3 by 3 size. The output layer comprises nodes, each node corresponding to a probability of belonging to a morphological class. Each morphological class corresponds to a morphology of the sperm cell being analyzed. It can for example be a known classification, the classes being: [0165] 0: undefined [0166] 1: normal [0167] 2: stump [0168] 3: distal droplet [0169] 4: decapitated [0170] 5: microcephalic [0171] 6: proximal droplet [0172] 7: DMR (distal midpiece reflex, usually referred to as curvature of the distal end of the midpiece) [0173] 8: head anomaly [0174] 9: tail anomaly: tail curved or wound [0175] 10: aggregates.

[0176] The output layer can thus include 11 classes.

[0177] The neural network is programmed in the Matlab environment (Matlab publisher: The Mathworks). The neural network was previously subjected to a learning (step 90), so as to parameterize the convolution filters. During the learning, learning sets were used, each set comprising: [0178] thumbnail images, extracted from images acquired by the image sensor; [0179] a manual annotation of the morphology of each sperm cell analyzed. The annotation was done on the basis of thumbnail images having been subjected to a holographic reconstruction.

Experimental Tests

[0180] Different tests were carried out to examine the performance levels of the detection neural network CNN.sub.d. The sensitivity of the detection was tested with respect to the application of a high-pass filter during the step 110. FIGS. 7A, 7B and 7C respectively represent an image acquired by the image sensor, a reference image, annotated manually, and an image resulting from the detection neural network. The image 7C was obtained without implementation of the filter during the step 110.

[0181] FIGS. 8A, 8B and 8C respectively represent an image acquired by the image sensor, a reference image, annotated manually, and an image resulting from the detection neural network. The image 8C was obtained by implementing a high-pass filter during the step 110.

[0182] It can be seen that the image 8C includes detected sperm cells, identified by arrows, which do not appear on the image 7C. The application of the filter enhances the detection efficiency of the classification neural network.

[0183] FIGS. 9A and 9B respectively represent the detection sensitivity and the detection specificity of the detection neural network CNN.sub.d (curves a, b and c), compared to a conventional algorithm (curve d). The conventional algorithm was based on image processing morphological operations of erosion/expansion type.

[0184] On each of the FIGS. 9A and 9B, the curves a, b and c correspond respectively: [0185] to the neural network trained with 10 000 annotations; [0186] to the neural network trained with 20 000 annotations; [0187] to the neural network trained with 20 000 annotations, each image acquired by the sensor having been subjected to a high-pass filtering.

[0188] The sensitivity and the specificity (ordinate axes) were determined on 14 different samples (abscissa axis) of diluted bovine semen (factor 10). It can be seen that the performance levels of the neural network, whatever the configuration (curves a, b and c), are greater than that of the conventional algorithm, which is noteworthy. Moreover, the best performance is obtained in the configuration c, according to which the neural network is trained with 20 000 annotations and is fed with an image having been subjected to a preprocessing with a high-pass filter.

[0189] It will be recalled that the sensitivity corresponds to the ratio of true positives to the sum of the ratios of true positives and false negatives, and that the specificity corresponds to the ratio of true positives to the sum of the ratios of true and false positives.

[0190] FIG. 10A is a confusion matrix representing the classification performance levels of a classification neural network CNN.sub.c parameterized to receive, as input image, an image resulting from a holographic reconstruction. The neural network is a convolutional neural network, as previously described, trained using images acquired using a defocused device, as previously described. By using this neural network, each image acquired is the subject of a holographic reconstruction in a sample plane extending through the sample, by using a holographic reconstruction algorithm as described in US20190101484. From the reconstructed image, a thumbnail image is extracted that is centered on the sperm cell to be characterized. The thumbnail image forms an input image of the neural network.

[0191] FIG. 10B is a confusion matrix representing the classification performance levels of a classification neural network parameterized to receive, as input image, a single thumbnail image, centered on the sperm cell to be characterized, and extracted from an image acquired according to a defocused configuration. The neural network is a convolutional neural network, structured as described in the step 160, except that the neural network is fed by a single thumbnail image.

[0192] FIG. 10C is a confusion matrix representing the classification performance levels of a classification neural network as described in the step 160, fed by a series of 5 thumbnail images extracted from images acquired in defocused configuration.

[0193] The axes of each confusion matrix correspond to the classes 1 to 10 previously described.

[0194] Note that the classification performance levels of the algorithms described in association with FIGS. 10B and 10C are greater than that described in FIG. 10A. For example, regarding the class 6 (proximal droplet anomaly), the classification performance level goes from 31% (FIG. 10A), to 61% (FIG. 10B), and to 96.9% (FIG. 10C).

[0195] Note also that a single input image makes it possible to obtain a satisfactory classification performance level. So, for a morphological classification, there is no need to have a series of images comprising several images. A single image can suffice. However, the classification performance level is better when several images are used.

[0196] The invention allows an analysis of a sample including sperm cells without recourse to a translation plate. Moreover, it allows a characterization directly from the image acquired by the image sensor, that is to say on the basis of diffraction figures of the different sperm cells, and without requiring recourse to numerical reconstruction algorithms. Currently, the recourse to such an algorithm is reflected by a processing time of 10 seconds per image. In other words 300 seconds for a series of 30 images. Moreover, as represented in FIGS. 10A to 10C, the invention allows a more accurate classification than a classification based on a neural network of identical structure, fed by reconstructed images.

[0197] Another advantage of the invention is the tolerance with respect to a defocusing. The inventors estimate that the invention tolerates offsets of 25 m between the optical system and the sample.