Method of stabilizing sonar images
20200400801 ยท 2020-12-24
Assignee
Inventors
Cpc classification
International classification
Abstract
A method of tracking a known object is presented, wherein a sonar image of an object which is distorted by an artifact associated with sonar imaging is compared with an image generated from a model of the object, and at least one of the two images is modified to reduce differences between them.
Claims
1. A method of tracking a known object, comprising: constructing a first image of the known object from received sonar data returned from the known object; comparing the first image to an second image generated from known characteristics of the known object; and modifying at least one of the first image and second image to reduce differences between the first image and the second image, wherein the differences between the first image and the second image arise from an artifact associated with sonar imaging.
2. The method of claim 1, wherein; the artifact arises from low resolution of a two dimensional sonar receiving device.
3. The method of claim 2, wherein; the artifact is a dilation.
4. The method of claim 1, wherein; the artifact is an erosion of the first image of a first area of the surface of the object.
5. The method of claim 1, wherein; the artifact at a voxel element of the first image arises from interference signals reflected from neighboring voxel elements of the object.
6. The method of claim 1, wherein; the artifact arises from interference signals from local multipath reflection.
7. The method of claim 1, wherein; the artifact arises from interference signals from multipath reflection from a surface different from the object removed from the surface of the object.
8. The method of claim 1, wherein; the second image is modified to more closely fit the first image to within a criterion.
9. The method of claim 8, wherein the criterion is that the second image is stable over a series of second images.
10. The method of claim 1, wherein; the first image is modified to more closely fit the second image to within a criterion.
11. The method of claim 10, wherein the criterion is that the second image is stable over a series of second images.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0011]
[0012]
[0013]
[0014]
[0015]
[0016]
[0017]
[0018]
[0019]
[0020]
[0021]
[0022]
[0023]
[0024]
[0025]
DETAILED DESCRIPTION OF THE INVENTION
[0026] It has long been known that data presented in visual form is much better understood by humans than data presented in the form of tables, charts, text, etc. However, even data presented visually as bar graphs, line graphs, maps, or topographic maps requires experience and training to interpret them. Humans can, however, immediately recognize and understand patterns in visual images which would be impossible for even the best and fastest computers to pick out. Much effort has thus been spent in turning data into images.
[0027] In particular, images which are generated from data which are not related to light are difficult to produce. One such type of data is sonar data, wherein a sonar signal is sent out from a generator into a volume of fluid, and the reflected sound energy from objects in the ensonified volume is recorded one or more detector elements. The term ensonified volume is known to one of skill in the art and is defined herein as being a volume of fluid through which sound waves are directed.
[0028] The sonar data from multielement detectors is generally recorded as points in three dimensional space as a function of range and of two orthogonal angles. These data in polar coordinate space are in turn generally reduced and presented as data from a three dimensional Cartesian coordinate space. The data may then be presented as height above the sea bed, for example, or depth below the surface, as a z coordinate, while the x and y coordinates could be chosen as west and north, for example. In other examples, the x or y coordinate could be chosen to be parallel to a wall or other long, mostly straight object.
[0029] One characteristic of sonar data is that it is very sparse, as the ensonified volume is generally water having only one or a few objects of interest. The volume of the fluid is generally divided into a series of cubes, and data is returned from a small percentage of the cubes. The resolution of the sonar is proportional to the linear dimension of the cubes, while the computation cost of recording the signal from each detector element and calculating from whence the signals have come is inversely proportional to the cube dimensions to the third power. There is then a tradeoff between resolution and computer power and time taken to produce an image from received data.
[0030] In other electromagnetic or ultra sound imaging technologies, the data are very dense. In an art unrelated to sonar imaging, medical imaging essentially has signals from each voxel, and the techniques for such imaging as CT scans, MRI scans, PET scans, and Ultrasound Imaging is not applicable to the sparse sonar data. In the same way, signals from sound waves sent out from the earths surface into the depths to return data of rock formations in the search for oil produce dense data, and techniques developed for such fields would not in general be known or used by one of skill in the art of sonar imaging.
[0031] The present invention is used to treat the sparse data from sonar imaging equipment to produce images which would be comparable to an optical image of an object in a sound wave transmitting medium if the object could in fact be seen through turbid water or other fluid or gas. These images are used to track and precisely place objects Optical inspection of objects in a fluid is often not possible because of smoke and fog in air, for example, or turbidity in water or other fluid. Sonar imaging of such objects is often used. However, if objects are to be placed, grasped, or moved in the fluid, a typical sonar image taken from a single point of view is not sufficient. The backside of the object is not viewable, nor is the background of the object in the sonar shadow viewable.
[0032]
[0033] As with optical holograms, images may be produced as would be seen from differing viewpoints. The inventor anticipates that a binocular image could be produced for display on a 3 dimensional display device for projecting the image to two eyes of a human observer.
[0034] When building a breakwater, the top (armor) layer is usually made with large heavy concrete blocks. These blocks must be placed sufficiently densely so as to minimize gaps between them to stop the egress of the underlying layers, and must be sufficiently heavy so as not to be moved by the action of waves and tides. Traditionally two layers of boulders, or in most cases cubic concrete blocks have been used. In order to reduce the amount of material required a new approach was introduced, where complex geometric volumes with overlapping parts were chosen. This allows only one layer of armor to be used while still meeting the minimum gap requirement. Photographs of typical blocks are shown in
[0035]
[0036] One advantage of the 3D visualization made possible by the 3D sonar detector is that the view point of the images drawn may be moved to take advantage of the human recognition of parallax to give the 3.sup.rd dimensional image information. As the Echoscope itself is fixed with respect to the scene, this virtual movement makes the shadowing effect more apparent. When the image shown from a viewpoint apart from the sonar array 16 as in
[0037] In order to show the backside of the block as the eyepoint is moved around, we obtain the sonar data on the relative coordinates of the surface of the block, and construct a model of the block in the computer as in
[0038] The model data image has now the same rotational orientation as the object, and appears to be the same distance away from the detector.
[0039] Many other methods of finding the best fit between sets of points in three dimensions could be used.
[0040] The ICP algorithm and other point matching algorithms require a time proportional to the number n of points in the first set of points to be matched times the number m of points in the second set of points. This time proportional to nm may be reduced to n log m by reducing the set of points from the model to just those points which could be seen from an Echoscope.
[0041]
[0042]
[0043] Before the first block in a set of blocks is laid, a sonar image of the background is recorded. The position and orientation of the sonar source and sonar imaging device are recorded, so that the background of the sonar shadow recalled from the recording, and can be filled in as the block is moved into place. The orientation of the block is known after it is placed, and the image of the block can be added to the background. As the blocks are placed, the position, orientation, etc. of each block is recorded so that the entire background may be matched. The measurement of the exact positions of the background blocks and the exact position of the equipment supporting the block being placed is at times not accurate enough to ensure correct placement of the blocks from positional data alone, and it is often preferable that the sonar background objects be measured as the block is being moved into position. As the block is being swung into place, the background is measured in the field of view int front of the swinging block. This background image is used by itself, or fit to a previously recorded background.
[0044] The block is moved into position to place it in a location and orientation with respect to the other blocks. The location and orientation must satisfy a criterion. One such criterion is that each block is supported by contact of at least three contact points with other blocks.
[0045] As the block is being moved and rotated, the movement and rotation is slow compared to the rate at which sonar images are recorded. The velocity and rotation of the block is measured by measuring the location of the excavator arm and the distance from the excavator arm to the block, and measuring the rotation of the block from ping to ping. The position and rotational velocity of the block is predicted at the time of the next ping, and and the previous set of points for matching model to sonar image is adjusted take into account the new position and rotation angle, so the iterative process of matching takes much less time, which allows us to track the block more accurately. For example, we anticipate that a set of points along one edge of the block can disappear from the sonar image, while another set of points on the opposite edge swings into view.
[0046] In viewing the block and background in the sonar image, the background can also be enhanced by using previously recorded orientations and positions to draw in the previously placed blocks. The sonar data is then much easier to understand, especially when the eyepoint is rotated back and forth to give enhanced 3D visualization. The previously recorded background orientations and positions may be augmented or replaced by images collected in as the blocks move into place.
[0047] A number of artifacts combine to produce sonar images which are quite distorted. When models are used to produce additional data for the sonar imaging visualization, the position and orientation of the model must be fit to the sonar data. Artifacts which distort the sonar image then affect the program which tries to match the sonar data points to the model data points, and different orientations of the model image with respect to the sonar image may give a fit to within the criterion chosen to end the iterative process. In particular, orientations chosen for each ping differ enough that the model image appears to jitter, even when the object is stationary.
[0048] Image artifacts arise, for example, due to the resolution of the sonar system. If the resolution of the system changes because the distance between the object and the Echoscope changes, the protrubences on an Accropode (a large concrete object used in underwater breakwaters to armor rip rap) may appear to be thicker than they should be because the diameter is measured at high resolution would have an uncertainty of a resolution element of 10 cm, and at low resolution of 30 cm.
[0049] Objects can appear smaller as at some angles the reflected energy is below the detection threshold. Consider a sphere. The surface normal of the center (surface) points directly at the sonar source and receiver, so reflects directly back at high intensity, which measurement is set to unity. The surface normal half way out to the edge of the sphere is indicates that 70% of the energy reflects back, while 30% is scattered more than 90 degrees to the incoming beam. The surface normal of the edge (surface) forms an angle of 90 degrees to the direction of the sonar beam, so reflects no energy directly back from that point. Setting the threshold for detection to 80% will show a sphere less than half the true size (even accounting for inflation). Another artifact of sonar imaging is sidelobe illumination. Every beam has 4 neighbors with a lower intensity, and so some energy from neighboring points of the surface will arrive at the detector and appear to come from another point. The beams can combine to show a surface where there is indeed a hole. Random data from other sound sources is an artifact which is very difficult to deal with if it is truly random, or even if it is not understood. Local Reflection/MultiPath effects is where a point on the object reflects sound onto another part of the object, which further reflects to the detector, and which causes points to appear in the wrong place.
[0050] The further away objects are, the less accurately we can track them. An Echoscope produces comparatively low resolution images compared to Images generated by light. A standard frequency (375 KHz) Echoscope has a resolution of 4848 elements (50 Degrees by 50 Degrees), giving an approximate angular resolution of 1 degree. However due to the way the image is constructed, we also have a limiting factor based on the physical size of the array (20 cm20 cm). This is known as Aperture size. The range above the point where 1 degree is greater than 20 cm is known as Far Field. The range below the point where 1 degree is less than 20 cm is known as Near Field. So the resolution of the standard frequency Echoscope is either 1 degree or 20 cm, whichever is greater.
[0051] (In Near Field you can make the aperture smaller by only using, say, 2424 elements. This gives a resolution of 2 degrees or 10 cm, whichever is greater. Limiting the number of elements to a 12 by 12 element array gives 5 cm resolution, etc.)
[0052] For 4848 resolution elements, or one degree resolution, the standard frequency Echoscope Far Field starts at around 11.5 m.
[0053]
[0054] The Accropode Image in
[0055] The Accropode Image in
[0056] If we place Accropode model data over the Accropode sonar image in
[0057] The Accropode in
[0058] A novel method for reducing this visual jittering effect has been implemented.
[0059] Since the cause of the distortion in the sonar image is known, we can preferably reverse the distortion of the sonar image. More preferably, we can distort the model image in a way which matches the distortion of the sonar image since we have more and more accurate data about the model than we do sonar data.
[0060] Preferable ways to match the distortion are dilation of the model image or erosion of the sonar image.
[0061] If we were to break the volume of interest into a volume of (say 1 cm cubes (voxels)), we can place the model inside this volume, if a 1 cm cube is predominantly inside the volume of the model we set that cube to be 1, otherwise it is 0. We can then inflate the model (by 1 cm), by looking at each cube, and its neighbors (the cubes that share any of its faces (there are 6 of these), or share any of its vertices (there are 26 of these)). If any of its neighbors have the value of 1, then set the value of this cube to 1. Every time we repeat this process the volume inflates by one voxel. Similarly if we wanted to make the data object smaller we could do some by Erosion. If we were to break the volume of interest into a volume of (say 1 cm cubes (voxels)), we can place the model inside this volume, if a 1 cm cube is predominantly inside the volume of the model we set that cube to be 1, otherwise it is 0. We can then deflate the model (by 1 cm), by looking at each cube, and its neighbors (the cubes that share any of its faces (there are 6 of these), or share any of its vertices (there are 26 of these)). If the voxel has a value of one and the number of its neighbors is less than or equal to some value (say one if we are looking at face neighbors), then we set that's voxel value to zero (removing it from the volume). Every time we repeat this process the volume deflates by one voxel.
[0062] Inflation, deflation, or scaling of the data takes less computer time than dilation or erosion, and is more preferable with limited computer equipment. Scaling of the model or sonar image data is one preferred embodiment of the invention. More preferably, inflation of the model data or deflation of the sonar data changes the fitting of the model to the sonar data better than scaling the data. The most preferable embodiment of the data is to inflate the model data to better fit the sonar data.
[0063] For simple (convex) objects, such as Cubes and spheres Inflation and scaling are the same. However this is not the case for more complex non-Convex objects.
[0064] There are many ways to inflate or deflate and scale objects. Preferable ways are based on Face Normals or Vertex Normals. Although Face Normals give a more uniform result, Vertex Normal technique as it is much simpler to implement and for our needs gives adequate results, and is the most preferred way to change the image of the model.
[0065]
[0066] The percentage inflation is of the model image is increased until the image of the Accropode is stable from ping to ping. As conditions (ie range) change, the percentage may be adjusted automatically or by hand.
[0067] Once the model Accropode image orientation and range have been determined, the sonar images may have the missing points drawn in. Or, the entire sonar image of the object may be replaced with an image of the model, and the image of the model can be drawn from any viewpoint at all. In particular, the model image may be used to guide the model with respect to either the sonar images or the model images of the background to a fit better than the resolution of sonar images.
[0068] Obviously, many modifications and variations of the present invention are possible in light of the above teachings. It is therefore to be understood that, within the scope of the appended claims, the invention may be practiced otherwise than as specifically described.