METHOD OF ESTIMATING THREE-DIMENSIONAL COORDINATE VALUE FOR EACH PIXEL OF TWO-DIMENSIONAL IMAGE, AND METHOD OF ESTIMATING AUTONOMOUS DRIVING INFORMATION USING THE SAME
20230143687 · 2023-05-11
Inventors
Cpc classification
G06V20/70
PHYSICS
G06V20/58
PHYSICS
G06V20/56
PHYSICS
International classification
G06V20/70
PHYSICS
Abstract
Proposed are a method of estimating a three-dimensional coordinate value for each pixel of a two-dimensional image, and a method of estimating autonomous driving information using the same, and more specifically, a method that can efficiently acquire information needed for autonomous driving using a mono camera. This method is able to acquire information having sufficient reliability in real-time without using expensive equipment such as a high-precision GPS receiver, a stereo camera or the like required for autonomous driving.
Claims
1. A method of estimating a three-dimensional coordinate value for each pixel of a two-dimensional image, the method comprising: a camera height input step of receiving height of a mono camera installed in parallel to ground; a reference value setting step of setting at least one among a vertical viewing angle, an azimuth angle, and a resolution of the mono camera; and a pixel coordinate estimation step of estimating a three-dimensional coordinate value for at least some of pixels with respect to ground of the two-dimensional image captured by the mono camera, based on the inputted height of the mono camera and a set reference value.
2. The method according to claim 1, wherein the pixel coordinate estimation step includes a modeling process of estimating the three-dimensional coordinate value by generating a three-dimensional point using a pinhole camera model.
3. The method according to claim 2, wherein the pixel coordinate estimation step further includes, after the modeling process, a lens distortion correction process of correcting distortion generated by a lens of the mono camera.
4. The method according to claim 1, further comprising, after the pixel coordinate estimation step, a non-corresponding pixel coordinate estimation step of estimating a three-dimensional coordinate value of a pixel that is not corresponding to the three-dimensional coordinate value among the pixels of the two-dimensional image from a pixel corresponding to the three-dimensional coordinate value using a linear interpolation method.
5. A method of estimating autonomous driving information using a method of estimating a three-dimensional coordinate value for each pixel of a two-dimensional image, the method comprising: a two-dimensional image acquisition step of acquiring the two-dimensional image captured by a mono camera; a coordinate system matching step of matching each pixel of the two-dimensional image and a three-dimensional coordinate system; and an object distance estimation step of estimating a distance to an object included in the two-dimensional image.
6. The method according to claim 5, wherein the coordinate system matching step includes the method of estimating a three-dimensional coordinate value for each pixel of a two-dimensional image of claim 4, and the object distance estimation step includes an object location calculation process of confirming the object included in the two-dimensional image, and estimating a direction and a distance to the object based on the three-dimensional coordinate value corresponding to each pixel.
7. The method according to claim 6, wherein at the object location calculation step, a distance to a corresponding object is estimated using a three-dimensional coordinate value corresponding to a pixel corresponding to the ground of the object included in the two-dimensional image.
8. A method of estimating autonomous driving information using a method of estimating a three-dimensional coordinate value for each pixel of a two-dimensional image, the method comprising: a two-dimensional image acquisition step of acquiring the two-dimensional image captured by a mono camera; a coordinate system matching step of matching each pixel of the two-dimensional image and a three-dimensional coordinate system; and a semantic information location estimation step of estimating a three-dimensional coordinate value of semantic information for autonomous driving included in the ground of the two-dimensional image.
9. The method according to claim 8, wherein the coordinate system matching step includes the method of estimating a three-dimensional coordinate value for each pixel of a two-dimensional image of claim 4, and further includes, after the semantic information location estimation step, a localization step of confirming a location of a corresponding vehicle on a HD-map for autonomous driving based on the three-dimensional coordinate value of semantic information for autonomous driving.
10. The method according to claim 9, wherein the localization step includes: a semantic information confirmation process of confirming corresponding semantic information for autonomous driving on the HD-map for autonomous driving; and a vehicle location confirmation process of confirming a current location of the vehicle on the HD-map for autonomous driving by applying a relative location with respect to the semantic information for autonomous driving.
11. The method according to claim 2, further comprising, after the pixel coordinate estimation step, a non-corresponding pixel coordinate estimation step of estimating a three-dimensional coordinate value of a pixel that is not corresponding to the three-dimensional coordinate value among the pixels of the two-dimensional image from a pixel corresponding to the three-dimensional coordinate value using a linear interpolation method.
12. The method according to claim 3, further comprising, after the pixel coordinate estimation step, a non-corresponding pixel coordinate estimation step of estimating a three-dimensional coordinate value of a pixel that is not corresponding to the three-dimensional coordinate value among the pixels of the two-dimensional image from a pixel corresponding to the three-dimensional coordinate value using a linear interpolation method.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0038]
[0039]
[0040]
[0041]
[0042]
[0043]
[0044]
[0045]
[0046]
[0047]
BEST MODE FOR CARRYING OUT THE INVENTION
[0048] Examples of a method of estimating a three-dimensional coordinate value for each pixel of a two-dimensional image, and a method of estimating autonomous driving information using the same according to the present invention may be diversely applied, and hereinafter, a most preferred embodiment will be described with reference to the accompanying drawings.
[0049]
[0050] Referring to
[0051] The camera height input step (S110) is a process of receiving the height (h) of a mono camera installed in parallel to the ground as shown in
[0052] The reference value setting step (S120) is a process of setting at least one among the vertical viewing angle (θ), azimuth angle (φ), and resolution of the mono camera as shown in
[0053] The pixel coordinate estimation step (S130) is a process of estimating a three-dimensional coordinate value for at least some of the pixels with respect to the ground of the two-dimensional image captured by the mono camera, based on the inputted height of the mono camera and a previously set reference value, and it will be described below in detail.
[0054] First, referring to
d=h/sin θ (Equation 1)
[0055] In addition, as shown in
[0056] For example, a three-dimensional point X, Y, and Z with respect to the ground may be expressed as shown in Equation 2 in terms of distance d, height h, vertical viewing angle θ, and the azimuth angle φ of the mono camera.
X=d cos θ sin Ø
Y=d cos θ cos Ø
Z=−h (Equation 2)
[0057] Thereafter, a three-dimensional coordinate value may be estimated by generating a three-dimensional point using a pinhole camera model.
[0058]
[0059] In addition, rotation matrix R for transforming the three-dimensional coordinate system of the mono camera's viewpoint into the coordinate system of a two-dimensional image may be expressed as shown in Equation 4.
R=R.sub.z(γ)R.sub.y(β)R.sub.x(α) (Equation 4)
[0060] Finally, in order to transform a point X, Y and Z of the three-dimensional coordinate system to a point of a two-dimensional image of the camera's viewpoint, the point of the three-dimensional coordinate system is multiplied by rotation matrix R as shown in Equation 5.
[0061] In this way, when the modeling process (S131) shown in
[0062] Generally, since a lens of a camera does not have a perfect curvature, distortion is generated in an image, and in order to estimate an accurate location, calibration for correcting the distortion is performed.
[0063] When external parameters of the mono camera are calculated through calibration of the mono camera, radial distortion coefficients k1, k2, k3, k4, k5 and k6 and tangential distortion coefficients p1 and p2 may be obtained.
[0064] The process as shown in Equation 6 is developed using the external parameters.
[0065] The relational equations of the image coordinate systems u and v obtained using the two points obtained before, focal lengths f.sub.x and f.sub.y, which are internal parameters of the mono camera, and principal points cx and cy are as shown in Equation 7.
u=f.sub.x*x″+c.sub.x
v=f.sub.y*y″+c.sub.y (Equation 7)
[0066] In the process as described above, when the height of the mono camera and the pinhole camera model are used, pixels and three-dimensional points corresponding to the ground may be calculated.
[0067] Hereinafter, the process described above will be described using an image actually captured by a mono camera.
[0068]
[0069] First,
[0070] Referring to
[0071] Here,
[0072] The data passing through the process may be used at an object location calculation step S151, a localization step S152, and the like, and this will be described below in more detail.
[0073]
[0074] Referring to
[0075] Describing in detail, a two-dimensional image captured by a mono camera is acquired at the two-dimensional image acquisition step (S210), and each pixel of the two-dimensional image and a three-dimensional coordinate system are matched at the coordinate system matching step (S220), and a distance to an object included in the two-dimensional image is estimated at the object distance estimation step (S230).
[0076] At this point, the coordinate system matching step (S220) may estimate a three-dimensional coordinate value for each pixel of the two-dimensional image through processes ‘S110’ to ‘S140’ of
[0077] Thereafter, at the object distance estimation step (S230), an object location calculation process of confirming an object (vehicle) included in the two-dimensional image as shown in
[0078] Specifically, at the object location calculation process, a distance to a corresponding object may be estimated using a three-dimensional coordinate value corresponding to a pixel corresponding to the ground (the ground on which the vehicle is located) of the object included in the two-dimensional image.
[0079]
[0080] In addition, the distance measured using LiDAR in the same situation is about 7.24 m as shown in
[0081]
[0082] Referring to
[0083] Describing in detail, a two-dimensional image captured by a mono camera is acquired at the two-dimensional image acquisition step (S310), and each pixel of the two-dimensional image and a three-dimensional coordinate system are matched at the coordinate system matching step (S320), and a three-dimensional coordinate value of semantic information for autonomous driving included in the ground of the two-dimensional image is estimated at the semantic information location estimation step (S330).
[0084] At this point, the coordinate system matching step (S320) may estimate a three-dimensional coordinate value for each pixel of the two-dimensional image through processes ‘S110’ to ‘S140’ of
[0085] In addition, after the semantic information location estimation step (S330), a localization step (S340) of confirming the location of a corresponding vehicle (a vehicle equipped with a mono camera) on a high-definition map (HD-map) for autonomous driving based on the three-dimensional coordinate value of the semantic information for autonomous driving may be further included.
[0086] Particularly, the localization step (S340) may perform a semantic information confirmation process of confirming corresponding semantic information for autonomous driving on the HD-map for autonomous driving, and a vehicle location confirmation process of confirming the current location of a vehicle on the HD-map for autonomous driving by applying a relative location with respect to the semantic information for autonomous driving.
[0087] In other words, as shown in
[0088] A method of estimating a three-dimensional coordinate value for each pixel of a two-dimensional image, and a method of estimating autonomous driving information using the same according to the present invention have been described above. It will be appreciated that those skilled in the art may implement the technical configuration of the present invention in other specific forms without changing the technical spirit or essential features of the present invention.
[0089] Therefore, it should be understood that the embodiments described above are illustrative and not restrictive in all respects.