Teaching device
12420402 ยท 2025-09-23
Assignee
Inventors
Cpc classification
B25J9/1664
PERFORMING OPERATIONS; TRANSPORTING
G05B19/42
PHYSICS
International classification
B25J13/08
PERFORMING OPERATIONS; TRANSPORTING
Abstract
A teaching device is provided with: a setting information storage unit for storing setting information defining a position and an attitude of a force sensor relative to a coordinate system set in a robot; and a virtual image superimposing and displaying unit for superimposing and displaying a virtual image in a real space including the robot or a prescribed object supporting the force sensor, or in a virtual space including a model of the robot or a model of the prescribed object, in such a way that a virtual image representing the force sensor adopts a position and an attitude corresponding to the setting information in the real space or the virtual space.
Claims
1. A teaching device, comprising: a setting information storage unit configured to store setting information that defines a position and a posture of a force sensor in a real space with respect to a coordinate system set in a real robot; and a virtual image superimposition display unit configured to superimpose and display a virtual image illustrating the force sensor on the real space or on an image of the real space, the real space including the real robot or a predetermined object that supports the force sensor, in such a way that the virtual image has a position and a posture according to the setting information in the real space.
2. The teaching device according to claim 1, wherein the virtual image is an image of a 3D model of the force sensor.
3. The teaching device according to claim 1, wherein the virtual image has asymmetry.
4. The teaching device according to claim 3, wherein the virtual image has rotational asymmetry about a predetermined coordinate axis in a coordinate system set in the force sensor.
5. The teaching device according to claim 1, wherein the virtual image includes an image illustrating a coordinate system set in the force sensor.
Description
BRIEF DESCRIPTION OF DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
DESCRIPTION OF EMBODIMENTS
(15) Next, embodiments of the present disclosure will be described with reference to drawings. A similar configuration portion or a similar functional portion is denoted by the same reference sign in the referred drawings. A scale is appropriately changed in the drawings in order to facilitate understanding. An aspect illustrated in the drawing is one example for implementing the present invention, and the present invention is not limited to the illustrated aspect.
First Embodiment
(16)
(17) The robot 10 includes a force sensor 21 mounted thereon, and is configured to be able to perform various types of work (such as fitting, pressing, phase alignment, polishing, and burr removal) by force control. For example, as illustrated in an enlarged view surrounded by a circle mark in
(18) An output value (force/moment) of the force sensor 21 is output as a value in a coordinate system 201 of the force sensor 21 as illustrated in
(19) In order to convert an output value of the force sensor 21 into a value in the coordinate system set in the robot 10, the robot system 100 holds, as internal data, setting information that defines a relationship between the coordinate system 201 of the force sensor 21 and the coordinate system set in the robot 10. The setting information that defines a relationship between the coordinate system 201 of the force sensor 21 and the coordinate system set in the robot 10 is set as a position and a posture of the force sensor 21 (coordinate system 201) with reference to the coordinate system set in the robot 10 as follows, for example.
Force sensor positional information(setting information)=(x1,y1,z1,th1,th2,th3)
Herein, x1, y1, and z1 are X, Y, Z coordinates of the force sensor 21 in the coordinate system of the robot 10, and th1, th2, and th3 are respectively rotational angles of the force sensor 21 about an X-axis, a Y-axis, and a Z-axis in the coordinate system of the robot 10.
(20) For example, the positional information about the force sensor 21 is set and input via a set value input interface of the teaching device 30 and is reflected in the robot controller 50. When there is an error in such positional information about the force sensor 21, the robot 10 may not operate as intended, but an operation setting of the robot 10 is extremely complicated, and thus it is generally difficult for a user to determine the cause (that there is an error in the positional information about the force sensor). Under such circumstances, as described below, the teaching device 30 according to the present embodiment superimposes and displays a virtual image illustrating the force sensor on a virtual space or a real space, based on setting information that defines a position and a posture of the force sensor, and thus a user can quickly and easily visually recognize an error when there is the error in the setting information.
(21)
(22) The teaching device 30 is connected to the robot controller 50 in a wired or wireless manner. The teaching device 30 is assumed to be a tablet type terminal herein, but various information processing devices such as a teaching operation panel, a smartphone, and a personal computer can be used as the teaching device 30.
(23)
(24) The teaching device 30 includes a setting information storage unit 301 that stores positional information (setting information) about the force sensor, a model data storage unit 302 that stores data about a 3D model of each object constituting the robot system 100 and data about a virtual image of the force sensor 21, a camera position posture estimation unit 303 that estimates a position and a posture of the camera 36, and a virtual image superimposition display unit 304. The setting information storage unit 301 may be formed of a non-volatile memory, or may be a temporary buffer formed in the RAM. Note that the setting information storage unit 301 may store various types of other information about a setting of the robot 10.
(25) The virtual image superimposition display unit 304 has a function of superimposing and displaying a virtual image illustrating the force sensor 21 on a real space including the robot 10 or a predetermined object that supports the force sensor 21, or a virtual space including a model of the robot 10 or a model of the predetermined object, in such a way that the virtual image illustrating the force sensor 21 has a position and a posture according to setting information in the real space or the virtual space. Thus, the virtual image superimposition display unit 304 includes an augmented reality image processing unit 305 having a function of generating an augmented reality image, or a virtual reality image processing unit 306 having a function of generating a virtual reality image.
(26) By the following procedure as an example, the camera position posture estimation unit 303 obtains a position and a posture of the camera 36 (teaching device 30) in a coordinate system (hereinafter, the coordinate system is assumed to be the robot coordinate system 101) fixed in a workspace of the robot 10, and tracks the position and the posture of the camera 36 (teaching device 30) in the coordinate system (robot coordinate system 101) fixed in the workspace. (A1) The camera position posture estimation unit 303 acquires an arrangement position of the robot 10 in the workspace (robot coordinate system 101) from a storage unit in the robot controller 50 or the teaching device 30. (A2) For example, the camera position posture estimation unit 303 captures a visual marker attached to the base portion of the robot 10 by the camera 36 (prompts a user to capture the visual marker). The visual marker is, for example, a marker known in the field having a visual pattern that can measure a position and a posture of a camera from a two-dimensional image in which the visual marker is captured. (A3) The camera position posture estimation unit 303 recognizes a position and a posture of the camera 36 (teaching device 30) in the robot coordinate system by image-processing an image in which the visual marker is captured, and registers the position and the posture of the camera 36 (teaching device 30) in the robot coordinate system 101. (A4) Hereinafter, the camera position posture estimation unit 303 obtains a movement of the camera 36 (teaching device 30) by the odometry technique, based on an output value of the inertial sensor 37, and continuously updates the position and the posture of the camera 36 (teaching device 30) in the robot coordinate system 101.
(27) The augmented reality image processing unit 305 has a function of superimposing and displaying, on an image (video) captured by the camera 36, a virtual image of the force sensor 21 in a position and a posture according to positional information (setting information) about the force sensor 21, based on a position and a posture of the camera 36 (teaching device 30) acquired from the camera position posture estimation unit 303.
(28) The virtual reality image processing unit 306 has a function of disposing a model of each object constituting the robot system 100 in a virtual space, based on actual arrangement positional information, and superimposing a virtual image of the force sensor 21 in the virtual space in a position and a posture according to positional information (setting information) about the force sensor 21.
(29) Two examples in which the virtual image superimposition display unit 304 displays a virtual image of the force sensor 21 according to positional information (setting information) about the force sensor 21 will be described below. Note that, in the following display examples, the positional information (setting information) about the force sensor 21 is assumed to be set as positional information with respect to the flange coordinate system 102 set in the robot 10.
Display Example 1
(30) As illustrated in
(31) Herein, it is assumed that the positional information (setting information) about the force sensor 21 is set to (dx, dy, dz2, 0, 0, 0), dx0, dy0, dz2dz by mistake. In this case, the robot controller 50 obtains force and moment in a point of application of force by regarding the force sensor 21 in a position as illustrated in
(32) In this way, by displaying, in an image in a real space in which the robot 10 is captured, the virtual image 21M of the force sensor 21 in a position and a posture according to setting information on the display unit 33, a user can instantly recognize an error in positional information (setting information) about the force sensor 21 from a comparison between the robot 10 and the position of the virtual image 21M on a display screen or a comparison between a real object of the force sensor 21 and the position of the virtual image 21M on the display screen.
(33) Note that, as illustrated in
Display Example 2
(34) As illustrated in
(35) Herein, it is assumed that the positional information (setting information) about the force sensor 21 is set to (Dx2, Dy2, Dz2, th1, th2, th3), Dx2Dx, Dy2Dy, Dz2Dz, th10, th2 #0, th30 by mistake. In this case, the robot controller 50 obtains force and moment in a point of application of force by regarding the force sensor 21 in a position as illustrated in
(36) In this way, by displaying, in an image in a real space in which the robot 10 is captured, the virtual image 21M of the force sensor 21 in a position and a posture according to setting information on the display unit 33, a user can instantly recognize an error in positional information (setting information) about the force sensor 21 from a comparison between the robot 10 (or the bracket 15A) and the position of the virtual image 21M or a comparison between a real object of the force sensor 21 and the position of the virtual image 21M.
(37) Note that, as illustrated in
(38) In Display Example 1 and Display Example 2 described above, it is an augmented reality in which the virtual image 21M of the force sensor 21 is superimposed and displayed on an image in a real space, but the teaching device 30 may display the image as illustrated in
(39) Note that, as a position of a point of view when an image by a virtual reality is generated, a camera position acquired by the camera position posture estimation unit 303 may be used, or a position of a point of view may be fixed in any position in a workspace. When a position of a point of view is fixed in any position in a workspace, the camera 36, the inertial sensor 37, and the camera position posture estimation unit 303 may be omitted from the teaching device 30.
Second Embodiment
(40) A second embodiment is an example when an arrangement position of a force sensor is different from that in the first embodiment described above.
(41) As illustrated in
(42) When the force sensor 22 is attached to the worktable 81 via the base 82, positional information (setting information) about the force sensor 22 is set as a position and a posture of coordinates in a robot coordinate system 101, for example. Herein, as an example, it is assumed that an original position and an original posture of the force sensor 22 are (x1, y1, z1, 0, 0, 0). Herein, it is assumed that the positional information (setting information) about the force sensor 21 is set to (x2, y2, z2, th1, th2, th3), x2x1, y2y1, z2z1, th10, th20, th30 by mistake. In this case, the robot controller 50 obtains force and moment in a point of application of force by regarding the force sensor 22 in a position as illustrated in
(43) In this way, by superimposing and displaying, in an image in a real space in which the worktable 81 (or the base 82) being a predetermined object that supports the force sensor 22, the robot 10, and the like are captured, the virtual image 22M of the force sensor 22 in a position and a posture according to setting information, a user can instantly recognize an error in positional information (setting information) about the force sensor 22 from a comparison between the worktable 81 (or the base 82) and the position of the virtual image 22M on a display screen or a comparison between a real object of the force sensor 22 and the position of the virtual image 22M on the display screen.
(44) Note that, as illustrated in
(45) In the display example described above, it is an augmented reality in which the virtual image 22M of the force sensor 22 is superimposed and displayed on an image in a real space, but the teaching device 30 may display the image as illustrated in
(46) As described above, according to each of the embodiments, a user can quickly and easily visually recognize an error when there is the error in a set position and a set posture of a force sensor from a virtual image of the force sensor being superimposed and displayed on a real space or a virtual space.
(47) Note that, in the first embodiment and the second embodiment described above, a 3D model of a force sensor and a coordinate system of the force sensor are illustrated as an example of a virtual image of the force sensor, but various aspects are possible as the virtual image of the force sensor. Herein, with reference to
(48) Specifically, the virtual image 23M has a shape being set in such a way that a connector 25 and a connector attachment portion 26 are provided on a side surface of a body portion 24 having a cylindrical shape. In this way, the virtual image 23M has a shape having rotational asymmetry also about the Z-axis. The virtual image 23M of the force sensor has such a shape having rotational asymmetry also about the Z-axis, and thus, even when there is an error in a posture (angle position) about the Z-axis of positional information (setting information) about the force sensor, the error can be easily recognized from the virtual image 23M of the force sensor.
(49) Note that, as illustrated in
(50) The present invention has been described above by using the typical embodiments, but it will be understood by those of ordinary skill in the art that changes, other various changes, omission, and addition may be made in each of the embodiments described above without departing from the scope of the present invention.
(51) In each of the embodiments described above, the example in which an augmented reality image or a virtual reality image is displayed on a display device of a teaching device as a tablet type terminal is illustrated, but such an augmented reality image or a virtual reality image may be displayed on a head-mounting display device, for example. When superimposition and display of a virtual image by an augmented reality are performed, a glasses-type AR display device that superimposes a virtual image on a real scene may be used.
(52) The functional block of the teaching device illustrated in
REFERENCE SIGNS LIST
(53) 10 Robot 11 Flange 15, 15A Bracket 21, 22 Force sensor 21M, 22M, 23M Virtual image 24 Body portion 25 Connector 26 Connector attachment portion 30 Teaching device 31 Processor 32 Memory 33 Display unit 34 Operating unit 35 Input/output interface 36 Camera 37 Inertial sensor 50 Robot controller 51 Processor 52 Memory 53 Input/output interface 54 Operating unit 81 Worktable 82 Base 100, 100A Robot system 101 Robot coordinate system 102 Flange coordinate system 201 Coordinate system 211, 212, 213 Mark 501 Storage unit 502 Operation control unit 301 Setting information storage unit 302 Model data storage unit 303 Camera position posture estimation unit 304 Virtual image superimposition display unit 305 Augmented reality image processing unit 306 Virtual reality image processing unit