DEVICE AND METHOD FOR CALIBRATING COORDINATE SYSTEM OF 3D CAMERA AND ROBOTIC ARM
20220080597 · 2022-03-17
Inventors
Cpc classification
B25J13/088
PERFORMING OPERATIONS; TRANSPORTING
G06T7/80
PHYSICS
International classification
Abstract
A calibration device fora 3D (three dimensional) camera and a robotic arm includes three plates with fixed relative positions and non-parallel separation disposed on a mount. Three spatial planes extending from the three plates intersect at a positioning point for external parameter correction. A calibration method for coordinates of a 3D (three dimensional) camera and a robotic arm using the calibration device is also specified in detail.
Claims
1. A calibration device for a 3D (three dimensional) camera and a robotic arm, comprising: three plates with fixed relative positions and non-parallel separation disposed on a mount; wherein three spatial planes extending from the three plates intersect at a positioning point for external parameter correction.
2. A calibration method for coordinates of a 3D (three dimensional) camera and a robotic arm, comprising: disposing a calibration device with three plates in a working environment of the robotic arm and a field of view of the 3D camera; capturing point cloud data of the calibration device with the 3D camera to generate a point cloud image; calculating a Z value difference and a vector angle of each set of adjacent point clouds according to the point cloud image; grouping the point cloud data according to the Z value difference and the vector angle to form point cloud groups; eliminating a set of point cloud groups with fewer point clouds to form three point cloud groups; establishing three plane equations by a least square method; calculating an intersection point of the three plane equations according to a three-plane intersection formula; and correcting external parameters by using the intersection point as a positioning point.
3. The method of claim 2, wherein the point cloud data is converted using a 3D information depth map with internal parameters of the 3D camera.
4. The method of claim 2, wherein the Z value difference is a difference between z-axis coordinates of the set of adjacent point clouds in a coordinate system of the 3D camera.
5. The method of claim 4, wherein the vector angle is an angle between a position vector formed by the set of adjacent point clouds and a horizontal or a vertical axis in the coordinate system of the 3D camera.
6. The method of claim 5, wherein the resolution of the point cloud image is set by the 3D camera, the slope of a plane of each plate is constant, differences between the z-axis coordinates of different sets of adjacent point clouds on the plane in the camera coordinate system are the same, the Z value differences of different sets of adjacent point clouds on the plane are the same, and the vector angles of different sets of adjacent point clouds on the plane are the same.
7. The method of claim 2, wherein the least square method comprises a plane equation:
8. The method of claim 2, wherein the three plane equation comprises:
9. The method of claim 2, further comprising: disposing the calibration device on the robotic arm; controlling the robotic arm to move the calibration device in the field of view of the 3D camera; obtaining positioning points relative to the camera coordinates through the 3D camera; and calibrating coordinates of the 3D camera and the robotic arm according to the positioning points.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0009]
[0010]
[0011]
[0012]
[0013]
[0014]
[0015]
[0016]
DETAILED DESCRIPTION
[0017] Please refer to
[0018] In
[0019] A 3D camera 17 is disposed in the working environment of the robotic arm 10. The 3D camera 17 has a camera coordinate system C. The information captured by the 3D camera 17 is transferred to the control device 16 for processing. However, the positional relationship of the camera coordinate system C with respect to the arm coordinate system R is unknown. The conversion and calibration between the camera coordinate system C and the arm coordinate system R are required to integrate the 3D camera 17 and the robotic arm 10.
[0020] To calibrate the robotic arm 10 with the 3D camera 17, firstly the calibration device 20 is disposed in the working environment of the robotic arm 10 and the calibration device 20 is within the field of view of the 3D camera 17. In the calibration device 20, three plates 22, 23 and 24 are disposed on the mount 21 in their relative positions and are separated from each other in unparalleled positions. The three spatial planes extending from the three plates 22, 23, 24 intersect at a single point as the exterior positioning point for parameter calibration.
[0021] The coordinates of the intersection point of the three spatial planes must be obtained through calculation. First, the 3D camera 17 can capture the 3D information of the calibration device 20, shown in
[0022] Taking the point cloud image of the plate 22 as an example, in
[0023] As shown in
[0024]
[0025] S1: Dispose the tool 14 on the robotic arm 10;
[0026] S2: Use the robotic arm 10 to move the tool 14 to any gesture, and use the force sensor of the robotic arm 10 to detect and record the tool gravity and the tool torque;
[0027] S3: Contact the center point of the tool 14;
[0028] S4: Detect the force change by the force sensor, and when the center point of the tool 14 is contacted, record the coordinates of the movable end 13 and record the contact force and the contact torque;
[0029] S5: Calculate the net contact force by subtracting the tool gravity from the recorded contact force, and subtract the tool torque from the recorded contact torque;
[0030] S6: Divide the net contact torque by the net contact force to obtain the net moment arm; and
[0031] S7: Calculate the coordinates of the center point of the tool 14 from the coordinates of the net moment arm and the movable end 13.
[0032] Using the point cloud groups on the planes 25, 27, and 28 of their respective plates 22, 23, and 24, the plane equations of the planes 25, 27, and 28 can be expressed using the conventional least square method, respectively. The plane equation of each plane can be expressed as:
[0033] The plane equations of the planes 25, 27, and 28 can be rewritten as:
a.sub.1x+b.sub.1y+c.sub.1z+d.sub.1=0
a.sub.2x+b.sub.2y+c.sub.2z+d.sub.2=0
a.sub.3x+b.sub.3y+c.sub.3z+d.sub.3=0
[0034] By calculating the planes intersection, the result can be obtained as:
wherein
and Det≠0
[0035] The present invention uses the most precise 3D information from the center part of the field of view of the 3D camera 17 to calculate the intersection point T of the three planes 25, 27, 28 in the camera coordinate system C as the positioning point T for calibrating the robotic arm 10. After the positioning point T is established, an example is given to illustrate the calibration method using the aforementioned positioning point T, but the method is not limited to the example.
[0036]
[0037]
[0038] M1: Dispose the calibration device 20 in a working environment of a robotic arm and a field of view of a 3D camera 17;
[0039] M2: Capture point cloud data of the calibration device with the 3D camera to generate a point cloud image;
[0040] M3: Calculate a Z value difference and the vector angle of each adjacent point cloud according to the point cloud image;
[0041] M4: Group the point cloud data according to the Z value differences and the vector angles to form point cloud groups;
[0042] M5: Eliminate a set of point cloud groups with fewer point clouds to form three point cloud groups;
[0043] M6: Establish three plane equations by the least square method;
[0044] M7: Calculate an intersection point of the three plane equations according to a three-plane intersection formula; and
[0045] M8: Correct external parameters by using the intersection point as a positioning point.
[0046] Since the embodiment uses the 3D information from the most precise part of the 3D camera to directly calculate the positioning point, it can reduce the error and cost of the instrument calibration. Also it can avoid deformation of the edge shape of the calibration device, so as to improve the precision of the positioning point.
[0047] Those skilled in the art will readily observe that numerous modifications and alterations of the device and method may be made while retaining the teachings of the invention. Accordingly, the above disclosure should be construed as limited only by the metes and bounds of the appended claims.