System, method and marker for the determination of the position of a movable object in space

11504855 · 2022-11-22

Assignee

Inventors

Cpc classification

International classification

Abstract

A system for determining the position of a movable object in space includes a marker which is to be applied to the object. The marker has a surface which is subdivided into a plurality of individual fields. The fields each have a statistical noise pattern. The system also includes an image capture unit which is remote from the object and is arranged to capture an image of the marker. The system further includes an image evaluation unit which stores a reference image of the noise patterns and is designed to locate at least one of the fields in the currently captured image of the marker by comparison with the reference image in order to determine a current position of the marker in space. There are corresponding methods for determining a position the object.

Claims

1. A system configured to determine a six dimensional position of an object that is movable with six degrees of freedom of movement in space, the system comprising: a marker configured to be applied to the object, the marker comprising a surface that is subdivided into a plurality of individual fields, each field having a statistical noise pattern; an image capture unit configured to capture an image of the object to provide an image of the marker, the image capture unit being remote from the object, the image capture unit comprising a plurality of cameras configured to observe the marker at different viewing angles; and an image evaluation unit configured to: a) store a reference image of the noise patterns; and b) locate a field in the image of the marker by comparing the image of the marker and the reference image to determine a current position of the marker in space.

2. The system of claim 1, wherein the fields are polygonal.

3. The system of claim 1, wherein the fields define rows and columns.

4. The system of claim 1, wherein the fields are circular or elliptical.

5. The system of claim 2, wherein the fields are rotated and/or shifted relative to one another.

6. The system of claim 1, wherein the noise patterns are different from one another.

7. The system of claim 1, wherein the noise patterns comprise broadband noise in a frequency domain.

8. The system of claim 1, wherein the noise patterns comprise grayscale patterns.

9. The system of claim 8, wherein a probability density of grayscale values within the noise patterns is a Gaussian distribution.

10. The system of claim 1, wherein the noise patterns comprise white noise.

11. The system of claim 1, wherein the image evaluation unit is configured to coarsely determine the position of the marker by detecting corners and/or edges of at least one member selected from the group consisting of the marker and the fields of the marker.

12. The system of claim 11, wherein the image evaluation unit is configured to finely determine the position of the marker by detecting the noise patterns.

13. The system of claim 1, wherein the image capture unit comprises a low-pass filter and/or a defocusing device.

14. The system of claim 1, wherein the noise patterns are generatable by subjecting white noise to low-pass filtering.

15. A method of determining a six dimensional position of an object that is movable with six degrees of freedom of movement in space, the method comprising: providing a marker comprising a surface that is subdivided into a plurality of individual fields, each field comprising a statistical noise pattern; storing a reference image of the noise patterns; applying the marker to the object; capturing an image of the marker on the object via a plurality of cameras which observe the markers at different viewing angles; and determining a position of the marker in space by comparing the image of the marker and the reference image, a field being located in the image of the marker.

16. The method of claim 15, wherein determining the current position of the marker comprises coarsely determining the position of the marker by detecting corners and/or edges of at least one member selected from the group consisting of the marker and the fields of the marker.

17. The method of claim 15, determining the current position of the marker comprises finely determining the position of the marker by detecting the noise patterns.

18. The method of claim 17, comprising using low-pass filtering or defocusing to capture the image of the noise patterns.

19. The method of claim 15, comprising subjecting white noise to low-pass filtering to generate the noise patterns.

20. The method of claim 15, comprising capturing the image of the marker from different viewing directions.

21. The method of claim 15, further comprising applying a plurality of markers to the object, and capturing image of the plurality of markers.

22. The method of claim 15, wherein the object comprises a movable machine part.

23. The method of claim 15, wherein the object comprises a movable component of an optical system.

Description

BRIEF DESCRIPTION OF THE DRAWINGS

(1) Exemplary embodiments of the disclosure are illustrated in the drawing and described in more detail below with reference thereto. In the figures:

(2) FIG. 1 shows an exemplary embodiment of a system for determining the position of a movable object in space, wherein the object is, by way of example, a robot arm of a machine, and a first exemplary embodiment of a marker for determining the position;

(3) FIG. 2 shows an exemplary image of a marker applied to the robot arm in FIG. 1, which image is captured by an image capture unit of the system in FIG. 1;

(4) FIG. 3 shows a block diagram illustrating an exemplary embodiment of a method for determining the position of a movable object in space via the system in FIG. 1;

(5) FIG. 4 shows a practical example of an individual noise pattern of the marker in FIG. 1;

(6) FIG. 5 shows another exemplary embodiment of a marker for determining the position of a movable object;

(7) FIG. 6 shows yet another exemplary embodiment of a marker for determining the position of a movable object;

(8) FIG. 7 shows yet another exemplary embodiment of a marker for determining the position of a movable object, wherein the marker is modified in comparison with FIG. 6;

(9) FIG. 8 shows an image of a marker, which is captured by an image capture unit of the system in FIG. 1, in the case of a grazing viewing direction of the image capture unit with respect to the marker; and

(10) FIG. 9 shows two noise patterns of a marker, wherein the left-hand noise pattern is unfiltered and the right-hand noise pattern has been subjected to low-pass filtering.

DESCRIPTION OF EXEMPLARY EMBODIMENTS

(11) FIG. 1 schematically shows a system for determining the position of a movable object 12 in space, which system is provided with the general reference sign 10. In the exemplary embodiment shown, the object 12 is a robot arm 14 of a machine 15. The robot arm 14 is movable, for example translationally movable and rotationally movable. The current location of the robot arm 14 can be expressed, for example, in Cartesian coordinates x, y, z, and the current orientation can be expressed as an angle of rotation about one or more of the axes x, y, z. The system 10 is fundamentally able to determine the position of the object 12 in space according to 6 degrees of freedom of movement, that is to say in a 6-dimensional manner. The working volume of the robot arm can be large, for example 1 m.sup.3 or more. FIG. 1 shows, by way of example, a coordinate system having x, y and z axes, wherein arrows 15x, 15y and 15z illustrate the degrees of freedom of rotation of the object 12 about the x, y and z axes.

(12) The system 10 has a marker 16 which is applied to the object 12. The marker 16 is stationary with respect to the object 12, that is to say it concomitantly moves with the object when the object 12 moves.

(13) The system 10 also has an image capture unit 18 remote from the object 12. The image capture unit 18 comprises a camera 20 which is, for example, a video camera and is equipped with an image sensor 22. The image sensor 22 may be in the form of a commercially available image sensor.

(14) The camera 20 is arranged to capture an image of the marker 16, wherein the camera regularly or continuously captures images of the marker 16 in chronological order, with the result that the changing positions of the object 12 can be permanently tracked. The camera 20 also has a lens 24 which images the marker 16 onto the image sensor 22.

(15) The camera 20 can be arranged in a manner fixed in space at least during operation of the system 10. However, the camera 20 may also be movable.

(16) The system 10 also has an image evaluation unit 26 which is connected to the image capture unit 18 and is used to evaluate the image of the marker 16, as captured by the camera 20, for the purpose of determining the current position of the marker 16 and therefore of the object 12.

(17) While the system 10 has been shown with only one camera 20 and one marker 16 in FIG. 1 for reasons of simplicity, it is understood that the system 10 may comprise a plurality of cameras 20 and a plurality of markers 16. In this case, the plurality of markers 16 can be applied to the object 12 at different positions and the cameras 20 can be distributed in space such that the cameras 20 observe the marker(s) 16 at different viewing angles.

(18) The system 10 can be in the form of a pure measurement system for tracking the movements of the object 12, but can also be used to control or regulate the movement of the object 12.

(19) Depending on the position of the object 12 in space, an image of the marker 16 having greater or less projection distortion is produced on the image sensor 22 of the camera 20. FIG. 2 shows, by way of example, such an image 28 of the marker 16 in FIG. 1 having projection distortion. The aim is now to determine the exact position of the object 12 from the image 28 of the marker 16 by determining the position of the marker 16 in the coordinate system of the camera 20. In the practical embodiment of the system 10, aspects such as camera calibration and aberrations of the camera 20 are taken into account, but this is not necessary for understanding the present disclosure. For understanding the disclosure, it is sufficient to consider the camera 20 as a pinhole camera.

(20) The result of the position determination by the image evaluation unit 26 is output by the image evaluation unit 26, as indicated with an arrow 27 in FIG. 1, for example to a display which is not illustrated or to a control or regulation system for controlling/regulating the movement of the object 12.

(21) In FIG. 1, the marker 16 is additionally illustrated on its own and in an enlarged form. “T” is used to denote the upper end of the marker 16, “L” is used to denote the left-hand end, “R” is used to denote the right-hand end and “B” is used to denote the lower end of the marker 16.

(22) The marker 16 is two-dimensional. The marker 16 is preferably flat. A surface 30 of the marker 16 is square in the exemplary embodiment shown in FIG. 1. Other surface shapes such as generally rectangular, generally polygonal, in particular multiangular, can likewise be considered. The surface 30 is subdivided into a plurality of fields 32, wherein a total of nine fields 32 of this type are present in the exemplary embodiment in FIG. 1. However, the number of fields 32 is not critical. It may be less than or greater than 9. The fields 32 are likewise square in the exemplary embodiment shown in FIG. 1. However, the fields 32 may also be generally rectangular or generally polygonal.

(23) The arrangement of the fields 32 is in the form of a matrix in rows and columns in the exemplary embodiment shown in FIG. 1, in which case the individual fields 32 each have the same orientation or rotational position with respect to one another.

(24) Each of the fields 32 is filled with a statistical noise pattern 34. In FIGS. 1, 2, 5 to 9, the noise patterns 34 are shown as regular patterns which are identical to one another for reasons of simplified illustration. However, it goes without saying that the noise patterns 34 have been generated by a stochastic process and therefore do not have any dominant periodicity and are also irregular. In addition, the noise patterns may be identical to one another. However, the noise patterns 34 may advantageously also be completely different from one another from field 32 to field 32. In other words, the noise patterns 34 can be unique with respect to one another. A practical implementation of an individual noise pattern 34 is shown in FIG. 4.

(25) The noise patterns 34 fill the surface 30 of the marker 16 as far as possible with a high degree of filling.

(26) The noise patterns 34 are preferably in the form of grayscale patterns, as shown by way of example in FIG. 4. The noise patterns 34 preferably have a bandwidth which is as high as possible in the (spatial) frequency domain. The noise patterns 34 are generated when producing the marker 16 and are stored as a reference image in the image evaluation unit 26.

(27) The cross-correlation function of all noise patterns 34 of the marker 16 is preferably as small as possible, that is to say the noise patterns 34 are preferably at least approximately uncorrelated from field to field 32.

(28) The noise patterns 34 preferably contain broadband white noise having a maximum possible signal value. The maximum signal value here means that the image sensor 22 of the camera 20 does not run into the black or white limit, but rather the grayscale range (intensity range) of the image sensor 22 is fully used. As already mentioned, the surface 30 of the marker 16 should be filled with the white noise to the maximum possible extent.

(29) White noise has a narrow autocorrelation function. The narrower the autocorrelation function, the better correlation-based algorithms can locate the noise patterns 34. White noise has a δ peak as the autocorrelation function. This is the narrowest autocorrelation function which is possible. White noise should therefore be given preference for correlation-based localization algorithms.

(30) Furthermore, the grayscale values of the noise patterns have a Gaussian distribution. A Gaussian probability density function (distribution of the grayscale values) has the maximum entropy and therefore the maximum information content for a given standard deviation. The given standard deviation relates to the grayscale range (intensity range) of the camera 20. The power of the noise can be selected, for example, in such a manner that the 3σ value of the Gaussian distribution fits into the grayscale value (intensity value) range of the camera.

(31) A noise pattern which is well-suited to correlation-based localization algorithms is not necessarily always but usually also suitable for other localization algorithms.

(32) An exemplary embodiment of the image evaluation unit 26 and a method of how the image evaluation unit 26 evaluates the image of the marker 16 currently captured by the camera 20 in order to determine the position of the marker 16 in space and therefore of the object 12 in space therefrom are described with reference to FIG. 3.

(33) The processes taking place in the image evaluation unit 26 are preferably subdivided into two superordinate process steps 36 and 38, wherein the process step 36 takes place first of all.

(34) The first process step 36 is a coarse determination which is used to coarsely locate the marker 16 and the fields 32 of the marker 16. The coarse determination is carried out on the basis of the currently captured image of the marker in the camera 20. This can be carried out, for example, by detecting edges 40 of the marker 16 and/or of the fields 32 of the marker 16, by detecting corners 42 of the marker 16 and/or of the fields 32 of the marker 16 and/or by calculating 44 a Radon/Hough transform. The steps 40, 42 and 44 can be efficiently implemented, for example, on an FPGA (field programmable gate array, that is to say an integrated digital technology circuit into which a logical circuit can be loaded). The items of information from the steps 40, 42 and 44 are linked to one another in an information fusion step 46 in order to obtain improved localization information therefrom still within the scope of the coarse determination of the position of the marker 16. The information fusion step 46 or a part of the latter can be carried out on a microprocessor in software. This is because the analyses of the edge detection 40, the corner detection 42 and the Radon/Hough transform result in a moderate number of coordinates of points and/or a list containing parameterized lines instead of a large quantity of image data.

(35) Coarse projection parameters and a coarse location of the fields 32 of the marker 16 result from the coarse determination.

(36) After the coarse determination 36, a fine determination 38 which is used to finely locate the fields 32, in particular, is carried out. The advantage of a coarse determination in advance of the fine determination is that the search areas for finely locating the fields 32 can be limited. This advantageously reduces the computing time.

(37) The fine determination 38 comprises a fine adjustment 48 of the noise patterns 34 of the marker 16 on the basis of the currently captured image of the marker 16 and the reference image of the noise patterns 34 which is stored in the image evaluation unit 26. The fine determination 38 results in precise projection parameters and precise locations of the fields 32 of the marker 16.

(38) Finally, in a step 50, the position of the marker 16 in space and therefore of the object 12 in space is calculated. The result is then output according to an arrow 52.

(39) Within the scope of the entire image evaluation or in a part of the image evaluation, it is also possible to use movement information from previous images of the marker 16, specifically in the steps 46, 48 and 50, as indicated with the reference sign 54. The movement information from previously captured images of the marker 16 can be used to reduce computing times by further limiting search areas or to improve the tracking of the movement of the object 12 through the system 10.

(40) If the position of the marker 16 is coarsely determined without evaluating the noise patterns 34, for example only on the basis of corner and/or edge detection of the marker 16 and/or of the fields 32, there is ambiguity with respect to a rotation of the marker 16 through 90° in each case in the exemplary embodiment of the marker 16. The cause of this is the strictly symmetrical and regular arrangement of the fields 32 having the same size and as squares with the same orientation.

(41) FIG. 5 shows an exemplary embodiment of a marker 16a in which this ambiguity is avoided. In the marker 16a, the fields 32a are all oriented differently with respect to one another. In the example shown with nine fields 32a, the fields 32a can be rotated through 9° with respect to one another from field to field, for example. This eliminates the ambiguity with respect to a 0° rotation of the marker 16a or with respect to rotations of the marker 16a through integer multiples of 90°. This makes it possible to unambiguously identify the orientation of the marker 16a already in the coarse determination phase according to FIG. 3 without evaluating the noise patterns 34. For example, the upper left-hand corner of the marker 16a can always be identified since the edges of the field 32a run parallel to the outer edges of the marker 16a only with respect to the left-hand upper field 32a.

(42) A further possible way (not illustrated) of avoiding ambiguities during the coarse determination via edge and corner detection is for the fields 32 of the marker 16 to have multiangular polygonal structures.

(43) Examples of how the fine determination 38 and therefore the precise determination of the position of the marker 16 in space can be carried out on the basis of an evaluation of the noise patterns 34 in the image of the marker 16 captured by the image capture unit 16 are described below.

(44) The fine determination is used to locate the fields 32 and is carried out by evaluating the noise patterns 34 in the camera image. The fine localization of the fields 32 can be implemented, for example, via an approach in which the parameter space (position parameters) is searched within the scope of a full search, that is to say completely. However, since there is prior knowledge from the coarse determination according to process step 36, the size of the search areas in the parameter space can be limited, which also limits the computing time.

(45) A full search means, in principle, processing the noise patterns 34 with a number of adopted parameter sets and therefore respectively generating an artificial camera image with the respectively adopted parameters. A similarity measure (merit function) is then calculated for each of the adopted (searched) parameter sets by comparing the reference image of the noise patterns (master patterns), which is stored in the image evaluation unit, with the corresponding currently captured image of the marker 16, wherein the master pattern is modified, distorted and/or shifted with the adopted parameters. That parameter set which achieves the maximum similarity (or the minimum non-similarity) between the modified/distorted/shifted master pattern and the currently captured noise pattern(s) 34 is the parameter set which ultimately finely locates the field(s) 32.

(46) Since a full search means that a plurality of parameter sets need to be tried in cascade loops, the calculation can be accelerated by virtue of a plurality of parallel units running on an FPGA, for example. At first glance, all computing units involve direct access to the captured image data relating to the marker 16. If a bottleneck occurs here, the image data can be copied/duplicated (for example in a plurality of block RAM units within an FPGA) before the computing units use the data. Consequently, the computing units which carry out the similarity calculation need not all access one memory.

(47) The (normalized) correlation coefficient can be used as a similarity measure for determining the parameter set which ultimately locates one or more or all of the fields 32. The normalized correlation coefficient also makes it possible to assess the current quality of the matching, that is to say how well/reliably a currently evaluated field has been located.

(48) During the full search, the maximum of the similarity measure is represented in order to find the appropriate parameter set representing the position of the field 32. The value of the maximum itself can be used as a measure of the quality with which a particular field is located. This value can also be transformed, for example, into a weighting factor which can be used in an estimation step when determining the position of the marker 16 in space. More reliably located fields 32 are taken into account more than more unreliable fields with this approach.

(49) Since there are a plurality of degrees of freedom for the full search, the calculation is still very complicated.

(50) A distorted coordinate is desirably calculated for each pixel within the noise pattern 34, for example using a calculation similar to the following calculation which is disclosed in the article by R. Hartley and A. Zisserman: Multiple View Geometry, Tutorial, CVPR, June 1999:

(51) x = x 1 x 3 = h 11 x + h 12 y + h 13 h 31 x + h 32 y + h 33 , y = x 2 x 3 = h 21 x + h 22 y + h 23 h 31 x + h 32 y + h 33

(52) There is one degree of freedom for the normalization in the above projection transformation equations. In this case, it is possible to investigate whether all parameters of the projection transformation have to be modified during the search. Such investigations can show that the practice of fixing some parameters (taking into account the result of the coarse determination) achieves sufficiently good results, while this greatly reduces the computing complexity that is involved.

(53) Another option is to investigate a linearization of the projection equation, which can result in a more efficient solution. This means that the coordinate of a projected point can be calculated more efficiently in a small search area around the coarsely determined parameters obtained from the coarse determination.

(54) Another possible way of reducing the computing complexity is to choose an approach in which the search is first of all carried out on an image of the marker 16 in the camera with a lower resolution, in which case the search is then refined on the same image with a higher resolution. The prior knowledge from the results of the lower-resolution image can be used to limit the search areas in the steps for locating the field(s) 32 in the higher-resolution image.

(55) According to another exemplary embodiment, instead of using the correlation coefficient which is calculated as a similarity measure on the basis of the image data, it is possible to choose an approach which is based on a census transform in order to accelerate the localization of the field(s) 32. The census transform is described, for example, in the article by Fridtjof Stein: Efficient Computation of Optical Flow Using the Census Transform, Proceedings of the 26th DAGM Symposium, 2004, pages 79-86.

(56) The fine determination (process step 38) for exactly locating the fields 32 of the marker 16 on the basis of the noise patterns 34 is particularly precise if, as already mentioned, the noise patterns have a high bandwidth in the frequency domain. The size of the bandwidth of the noise patterns directly determines the accuracy of the system 10. Noise patterns with white noise have a narrow autocorrelation function, ideally a 2-dimensional delta pulse, and are therefore optimally suitable for correlation-based block matching methods in order to locate the fields 32.

(57) Noise patterns having full-frequency, that is to say very broadband, content enable a good estimation of blurring of the image, which is caused by movement, for speed estimations. That is to say, the frequency content of the image is determined only by the imaging behavior of the camera and by movement-related blurring.

(58) If the fields 32 have straight lines as an edge, as shown in FIG. 1, an efficient coarse preliminary estimation of the projection parameters and marker position candidates is enabled. An ellipse fit, as is desired for circular fields, is not required here. Instead, it is possible to use Radon/Hough transform-type analyses in order to find the lines and field position candidates. Additionally or alternatively, the corners can be analyzed in order to find the frames or edges of the field(s). Mathematically speaking and from the point of view of accuracy, an ellipse fit may be better, but Radon/Hough transform-type analyses and/or corner detections can be efficiently implemented in hardware/FPGA. A Radon transform calculation can be efficiently carried out by considering the so-called “central slice theorem” or “projection slice theorem”.

(59) According to another exemplary embodiment of a marker 16b, which is shown in FIG. 6, the marker 16b may have circular fields 32b into which the surface 30b of the marker 16b is subdivided. Like in the marker 16 and the marker 16a, the fields 32b are each filled with a statistical noise pattern 34b, in which case reference can be made to the description of the noise patterns 34 of the marker 16 with regard to the configuration of the noise patterns 34b.

(60) In the case of the marker 16b, the high accuracy with which the position of the marker 16b is determined can be achieved on account of the noise patterns 34b with a high bandwidth. The use of circular or else elliptical fields 32b can be advantageous since, even with projective distortion, a circular shape results in a circular/elliptical shape, whereas the high measurement accuracy is achieved by the statistical noise patterns within the fields 32b.

(61) The position and/or orientation of the marker 16b can be coarsely determined via standard circle or ellipse detection methods. The detected ellipse parameters or the corresponding list of coordinates can be used to calculate the position in addition to the results from the fine localization of the noise patterns 34b.

(62) In the case of the marker 16c having circular (or elliptical) fields 32b, the surface occupancy or the degree of filling of the surface 30c of the marker 16c can be increased according to the exemplary embodiment of a marker 16c in FIG. 7 by not strictly arranging the fields 32c in rows and columns, but rather offsetting them with respect to one another in such a manner that the highest possible surface occupancy of the surface 30c of the marker 16c is achieved. In the arrangement according to FIG. 6, twelve fields 32c, for example, find space on the surface 30c, instead of only nine fields 32b, with the same size of the fields 32c and 32b.

(63) The increase in the degree of filling of the surface 30c and therefore the increase in the number of fields which are filled with statistical noise patterns means that there are more observation pixels. An increase in the number of observation pixels in the fields reduces the standard deviation when locating the fields 32c.

(64) A further aspect is described below with reference to FIGS. 8 and 9.

(65) FIG. 8 shows an image 28′, comparable to the image 28 in FIG. 2, of the marker 16 on the object 12, but the camera 20 now captures the marker 16 with strongly grazing incidence of the viewing direction. The result is that the image 28′ of the marker 16 has severe projection distortion on the image sensor 22 of the camera 20. This means that at least parts of the image 28′ are sampled at a lower spatial sampling rate.

(66) This reduced sampling rate on account of severe projection distortion of the marker 16 results in so-called “aliasing”, in particular when the bandwidth of the frequency of the noise in the noise patterns is maximized, as is preferred within the scope of the present disclosure. Aliasing is an effect and, in the field of signal analysis, denotes errors which occur when frequency components which are higher than half the sampling frequency occur in the signal to be sampled. Aliasing results in poor or even completely incorrect localization or position determination of the fields 32 in the image 28′ and therefore in an unusable position determination of the marker 16.

(67) Since aliasing occurs already during image formation, it cannot be reversed in the image evaluation unit 26. However, this can be addressed in the structure of the system 10. The bandwidth of the noise patterns is desirably limited, with the result that aliasing does not occur for projection distortions which are not too severe.

(68) For very severe projection distortions, the position measurement results are very uncertain—at least for some degrees of freedom. The threshold for what is “excessively severe” projection distortion is desirably determined during system development. This can be addressed, for example, by virtue of the image capture unit 18 having a plurality of cameras 20 which observe the marker 16 in different viewing directions, and/or a plurality of markers 16 can be applied to the object 12 in a distributed manner. If there are a plurality of cameras, they should be arranged in such a manner that at least one camera always observes a noise pattern 34 with a sufficiently high sampling rate, with the result that aliasing is avoided or is small enough. If a plurality of markers 16 are used, they can be arranged in such a manner that at least one of the markers 16 is observed by the camera 20 without excessively great projection distortion.

(69) Another possible way of avoiding aliasing is to limit the frequency band (in principle an anti-aliasing filter). This can be achieved, for example, by virtue of the optics of the camera 20 (lens 24) being deliberately defocused. Another possibility is to generate the noise patterns 34 by subjecting white noise to low-pass filtering.

(70) The band limitation by subjecting the noise pattern to low-pass filtering during its generation is advantageous (for reducing aliasing effects). This is because, in this procedure, there is better control over the filter, which is applied to the noise pattern 34, in comparison with the filtering via optics adjustment while recording the marker 16 using the camera 20. Control over the filter, which is applied to the noise pattern during its generation, is advantageous since the filter applied to the noise pattern determines the autocorrelation function of the resulting noise pattern. A good trade-off between sufficient low-pass filtering and the maintenance of a narrow autocorrelation function can be achieved thereby.

(71) FIG. 9 shows, on the left, an example of a statistical noise pattern 34 without low-pass filtering during generation, whereas FIG. 8 shows, on the right, the noise pattern 34 with low-pass filtering. In this case, a 2D binomial filter, for example, can be used to smooth the noise pattern 34.

(72) An alternative filter may be a filter which results in a rotationally symmetrical autocorrelation function e∧(−λ*sqrt(x2+y2)). λ determines the “narrowness” of the autocorrelation function which relates to the bandwidth in the frequency domain. The coordinates x and y are the spatial coordinates in the image plane.

(73) Anti-aliasing can be directed not only to the noise patterns 34, but also to the edges of the fields 32 and/or the edges of the marker 16 itself, since aliasing effects can also occur there.