METHOD AND APPARATUS FOR GENERATING AN INITIAL SUPERPIXEL LABEL MAP FOR AN IMAGE
20180005039 · 2018-01-04
Inventors
Cpc classification
G06V20/46
PHYSICS
International classification
Abstract
A method and an apparatus for generating an initial superpixel label map for a current image from an image sequence are described. The apparatus includes a feature detector that determines features in the current image. A feature tracker then tracks the determined features back into a previous image. Based on the tracked features a transformer transforms a superpixel label map associated to the previous image into an initial superpixel label map for the current image.
Claims
1. A method for generating an initial superpixel label map for a current image from an image sequence, the method comprising: determining features in the current image; tracking the determined features back into a previous image; and transforming a superpixel label map associated to the previous image into an initial superpixel label map for the current image based on the tracked features, the method further comprising adding features at the borders of the current image.
2. The method according to claim 1, further comprising generating meshes consisting of triangles for the current image and the previous image from the determined features.
3. The method according to claim 2, further comprising determining for each triangle in the current image a transformation matrix of an affine transformation for transforming the triangle into a corresponding triangle in the previous image.
4. The method according to claim 3, further comprising transforming coordinates of each pixel in the current image into transformed coordinates in the previous image using the determined transformation matrices.
5. The method according to claim 4, further comprising initializing the superpixel label map for the current image at each pixel position with a label of the label map associated to the previous image at the corresponding transformed pixel position.
6. The method according to claim 4, further comprising clipping the transformed coordinates to a nearest valid pixel position.
7. (canceled)
8. The method according to claim 1, further comprising assigning a pixel split-off from a main mass of a superpixel in the initial superpixel label map to a neighboring superpixel.
9. A computer readable storage medium having stored therein instructions for generating an initial superpixel label map for a current image from an image sequence, which when executed by a computer, cause the computer to: determine features in the current image; track the determined features back into a previous image; and transform a superpixel label map associated to the previous image into an initial superpixel label map for the current image based on the tracked features, said instructions further causing the computer to add features at the borders of the current image.
10. An apparatus for generating an initial superpixel label map for a current image from an image sequence, the apparatus comprising: a feature detector configured to determine features in the current image; a feature tracker configured to track the determined features back into a previous image; and a transformer configured to transform a superpixel label map associated to the previous image into an initial superpixel label map for the current image based on the tracked features said apparatus being configured to add features at the borders of the current image.
11. An apparatus for generating an initial superpixel label map for a current image from an image sequence, the apparatus comprising a processing device and a memory device having stored therein instructions, which, when executed by the processing device, cause the apparatus to: determine features in the current image; track the determined features back into a previous image; and transform a superpixel label map associated to the previous image into an initial superpixel label map for the current image based on the tracked features said instructions, when executed by the processing device, further causing the apparatus to add features at the borders of the current image.
12. The method according to claim 1, further comprising adding said added features at each corner and at the center of each border of the current image.
13. The computer readable storage medium according to claim 9, wherein said instructions further cause the computer to add said added features at each corner and at the center of each border of the current image.
14. The apparatus for generating an initial superpixel label map according to claim 10, wherein said apparatus is configured to add said added features at each corner and at the center of each border of the current image.
15. An apparatus for generating an initial superpixel label map according to claim 11, wherein said apparatus is configured to add said added features at each corner and at the center of each border of the current image.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0029]
[0030]
[0031]
[0032]
[0033]
[0034]
[0035]
[0036]
[0037]
[0038]
[0039]
[0040]
DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
[0041] For a better understanding the principles of some embodiments shall now be explained in more detail in the following description with reference to the figures. It is understood that the proposed solution is not limited to these exemplary embodiments and that specified features can also expediently be combined and/or modified without departing from the scope of the present principles as defined in the appended claims.
[0042] The present approach for a fast label propagation is visualized in
[0043] The Matrix elements t.sub.1,j to t.sub.4,i determine the rotation, shearing, and scaling, whereas the elements t.sub.5,i to t.sub.6,i determine the translation. Using this transformation matrix of the triangle the homogeneous coordinates of each pixel (x,y,1).sub.k+1.sup.T in frame k+1 can be transformed into coordinates ({tilde over (x)},{tilde over (y)},1).sub.k.sup.T; of frame k:
[0044] The coordinates are clipped to the nearest valid pixel position. These are used to lookup the label in the superpixel label map of frame k, which is shown in
[0045] Occasionally after the warping some pixels are split-off from the main mass of a superpixel due to the transformation. As the spatial coherency of the superpixels has to be ensured, these fractions are identified and assigned to a directly neighboring superpixel. As this step is also necessary if a dense optical flow is used it does not produce additional computational overhead.
[0046] To analyze the performance of the proposed approach some benchmark measurements have been performed. The results are presented in
[0051] From the figures it can be seen that the proposed mesh-based propagation method produces a comparable segmentation error while the average temporal length is only slightly decreased. While the 2D boundary recall stays the same for the approach TSP w/mesh it is even improved for the approach TCS w/mesh.
[0052] In order to evaluate the runtime performance improvements in terms of computational costs the average runtime of the dense optical flow based label propagation and the mesh-based propagation was measured. Thereby, the label propagation method that is used in the original versions of TSP and TCS as well as a Horn&Schunck implementation is used as a reference. The performance benchmarks were done on an Intel i7-3770K @ 3.50 GHz with 32 GB of RAM. The results are summarized in Table 1.
[0053] From Table 1 it can be seen that the proposed method performs the superpixel label propagation task more than 100 times faster than the originally proposed methods while creating nearly the same segmentation quality as seen in
TABLE-US-00001 TABLE 1 Average runtime needed to propagate a superpixel label map onto a new frame Label propagation method Avgerage time/frame Method used in TSP and TCS 814.9 ms Horn&Schunck 114.3 ms Proposed approach 6.1 ms
[0054]
[0055] One embodiment of an apparatus 20 for generating an initial superpixel label map for a current image from an image sequence according to the present principles is schematically depicted in
[0056] Another embodiment of an apparatus 30 for generating an initial superpixel label map for a current image from an image sequence according to the present principles is schematically illustrated in
[0057] For example, the processing device 31 can be a processor adapted to perform the steps according to one of the described methods. In an embodiment said adaptation comprises that the processor is configured, e.g. programmed, to perform steps according to one of the described methods.
[0058] A processor as used herein may include one or more processing units, such as microprocessors, digital signal processors, or combination thereof.
[0059] The local storage unit 22 and the memory device 32 may include volatile and/or non-volatile memory regions and storage devices such hard disk drives and DVD drives. A part of the memory is a non-transitory program storage device readable by the processing device 31, tangibly embodying a program of instructions executable by the processing device 31 to perform program steps as described herein according to the present principles.
REFERENCES
[0060] [1] Ren et al.: “Learning a classification model for segmentation”, IEEE International Conference on Computer Vision (ICCV) (2003), pp. 10-17. [0061] [2] Munderloh et al.: “Mesh-based global motion compensation for robust mosaicking and detection of moving objects in aerial surveillance”, IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR) (2011), 1st Workshop of Aerial Video Processing (WAVP), pp. 1-6. [0062] [3] Chang et al.: “A Video Representation Using Temporal Superpixels”, IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2013), pp. 2051-2058. [0063] [4] Reso et al.: “Superpixels for Video Content Using a Contour-based EM Optimization”, Asian Conference on Computer Vision (ACCV) (2014), pp. 1-16. [0064] [5] Shi et al.: “Good features to track”, IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (1994), pp. 593-600. [0065] [6] Xu et al.: “Streaming Hierarchical Video Segmentation”, European Conference on Computer Vision (ECCV) (2012), pp. 1-14. [0066] [7] Van den Bergh et al.: “Online Video SEEDS for Temporal Window Objectness”, IEEE International Conference on Computer Vision (ICCV) (2013), pp. 377-384. [0067] [8] Horn et al.: “Determining optical flow”, Artificial Intelligence, Vol. 17 (1981), pp. 185-203.