Incorporating Vision System and In-Hand Object Location System for Object Manipulation and Training
20210023715 ยท 2021-01-28
Assignee
Inventors
- Biao Zhang (West Hartford, CT)
- Yixin Liu (South Windsor, CT, US)
- Thomas A. Fuhlbrigge (Ellington, CT, US)
- Saumya Sharma (Albany, NY, US)
Cpc classification
B25J9/163
PERFORMING OPERATIONS; TRANSPORTING
International classification
Abstract
A system and method of object manipulation and training including providing at least one robotic hand including a plurality of grippers connected to a body and providing a plurality of cameras disposed in a periphery surface of the grippers. The method also includes providing a plurality of tactile sensors disposed in the periphery surface of the grippers and actuating the grippers to grasp an object. The method further includes detecting a position of the object with respect to the robotic hand via a first image feed from the tactile sensors and detecting a position of the object with respect to the robotic hand via a second image feed from the cameras. The method also includes generating instructions to grip and manipulate an orientation of the object based on the first and the second image feeds for a visualization of the object relative to the robotic hand.
Claims
1. A method of object manipulation and training, comprising: providing at least one robotic hand including a plurality of grippers connected to a body; providing a plurality of cameras disposed in a periphery surface of the plurality of grippers; providing a plurality of tactile sensors disposed in the periphery surface of the plurality of grippers; actuating the plurality of grippers to grasp an object; detecting a position of the object with respect to the at least one robotic hand via a first image feed from the plurality of tactile sensors; detecting a position of the object with respect to the at least one robotic hand via a second image feed from the plurality of cameras; and generating instructions to grip and manipulate an orientation of the object based on the first and the second image feeds for a visualization of the object relative to the at least one robotic hand, wherein the at least one robotic hand, the plurality of grippers, the plurality of cameras and the plurality of tactile sensors are electrically connected to a controller.
2. The method of claim 1, wherein the plurality of cameras each include a fish eye lens and is disposed in the body of the at least one robotic hand.
3. The method of claim 1, further comprising providing at least one illumination surface disposed on the peripheral surface of the plurality of grippers.
4. The method of claim 3, wherein the at least one illumination surface is a pressure-activated luminescent surface.
5. The method of claim 1, wherein the plurality of grippers include mechanical linkages connecting the plurality of grippers to the body of the at least one robotic hand.
6. The method of claim 5, wherein the mechanical linkages include actuators configured to provide motion to the plurality of grippers via the controller.
7. The method of claim 3, wherein the at least one illumination surface is configured to provide a light source for the plurality of cameras.
8. The method of claim 1, wherein the controller comprises a tactile sensor array electrically connected to the plurality of tactile sensors, a vision array electrically connected to the plurality of cameras, an acute actuator control module and a gross actuator control module connected to the at least one robotic hand to move the plurality of grippers, and a central controller configured to connect to and to control each component via a communication bus.
9. The method of claim 1, wherein each of the plurality of tactile sensors comprises a reflective film sandwiched between at least two tactile layers, a light source and a camera.
10. The method of claim 9, wherein the at least two tactile layers are elastomers.
11. The method of claim 9, wherein the camera and the light source are disposed adjacent only one of the at least two tactile layers, and wherein the light source and the camera are electrically connected to the controller to render a 3D image of a touched surface by the plurality of tactile sensors.
12. The method of claim 1, further comprising: performing a pick procedure on the object based on the generated instructions; determining whether or not the image feeds from the visualization of the object correlates with the generated instructions; correcting the gripping and manipulating of the object based on the determining; and placing the object in an assembly of parts.
13. The method of claim 12, wherein if the correcting fails, then dropping the object and performing a re-pick of the object.
14. A robotic hand, comprising: a plurality of grippers and a body; a plurality of cameras disposed in a peripheral surface of the plurality of grippers; at least one illumination surface disposed on a periphery surface of the plurality of grippers; and a plurality of tactile sensors disposed in the peripheral surface of the plurality of grippers, wherein the at least one robotic hand, the plurality of grippers, the plurality of cameras, the at least one illumination surface and the plurality of tactile sensors are electrically connected to a controller.
15. The robotic hand device of claim 13, wherein the at least one illumination surface is a pressure-activated luminescent surface.
16. The robotic hand device of claim 13, wherein the plurality of grippers include mechanical linkages connecting the plurality of grippers to the body.
17. The robotic hand device of claim 15, wherein the mechanical linkages include actuators configured to provide motion to the plurality of grippers via the controller.
18. The robotic hand device of claim 13, wherein the at least one illumination surface is configured to provide a light source for the plurality of cameras.
19. A non-transitory computer-readable medium storing instructions that, when executed by a processor of a computer, cause the processor to perform operations comprising: actuating the plurality of grippers to grasp an object; detecting a position of the object with respect to the at least one robotic hand via a first image feed from the plurality of tactile sensors; detecting a position of the object with respect to the at least one robotic hand via a second image feed from the plurality of cameras; and generating instructions to grip and manipulate an orientation of the object based on the first and the second image feeds for a visualization of the object relative to the at least one robotic hand.
20. The operations of claim 19, further comprising: performing a pick procedure on the object based on the generated instructions; determining whether or not the image feeds from the visualization of the object correlates with the generated instructions; correcting the gripping and manipulating of the object based on the determining; and placing the object in an assembly of parts.
Description
BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWING(S)
[0010]
[0011]
[0012]
[0013]
[0014]
[0015]
[0016]
[0017]
[0018]
[0019]
[0020]
[0021]
[0022]
[0023]
[0024]
[0025]
DETAILED DESCRIPTION OF THE INVENTION
[0026] All references, including publications, patent applications, and patents, cited herein are hereby incorporated by reference to the same extent as if each reference were individually and specifically indicated to be incorporated by reference and were set forth in its entirety herein.
[0027] The use of the terms a and an and the and at least one and similar referents in the context of describing the invention (especially in the context of the following claims) are to be construed to cover both the singular and the plural, unless otherwise indicated herein or clearly contradicted by context. The use of the term at least one followed by a list of one or more items (for example, at least one of A and B) is to be construed to mean one item selected from the listed items (A or B) or any combination of two or more of the listed items (A and B), unless otherwise indicated herein or clearly contradicted by context. The terms comprising, having, including, and containing are to be construed as open-ended terms (i.e., meaning including, but not limited to,) unless otherwise noted. Recitation of ranges of values herein are merely intended to serve as a shorthand method of referring individually to each separate value falling within the range, unless otherwise indicated herein, and each separate value is incorporated into the specification as if it were individually recited herein. All methods described herein can be performed in any suitable order unless otherwise indicated herein or otherwise clearly contradicted by context. The use of any and all examples, or exemplary language (e.g., such as) provided herein, is intended merely to better illuminate the invention and does not pose a limitation on the scope of the invention unless otherwise claimed. No language in the specification should be construed as indicating any non-claimed element as essential to the practice of the invention.
[0028] The invention particularly describes the use of computer-aided design (CAD) model/synthetic data of the objects being handled/assembled, together with the tactile imaging information with reference to a robotic hand, and the robot joint coordinate information that is easily accessible as well. This pool of information can allow coordinated movement and easy manipulation of the object which is being picked or assembled. This pool of information may also allow for easier forecasting of robot gestures or grasp planning.
[0029] Referring now to
[0030] Referring now to
[0031] Referring now to
[0032] Referring now to
[0033] Further, the in-hand sensor 40 may include a block of transparent rubber or gel, one face of which is coated with metallic paint. When the paint-coated face is pressed against an object, it conforms to the object's shape. The metallic paint makes the object's surface reflective, so its geometry becomes much easier for computer vision algorithms to infer. Mounted on the sensor opposite the paint-coated face of the rubber block are colored lights/LEDs 50a-d and a single camera 45. This system needs to have colored lights at different angles, and then it has the reflective material, and by looking at the colors, a computer can figure out a 3-D shape of what is being sensed or touched.
[0034] Referring now to
[0035] In some embodiments, an in-hand object location system may be used to determine the location of a part held within a robotic hand. This system may additionally provide information about the geometry of the object 90. This system may also be used to find a different location that may provide a better grasp of the object 90. Such an in-hand object location system requires a light source and a detector or camera unit within the robotic hand. Mounting an elaborate light source while maintaining a compact robot hand/fingers may be challenging, but the tactile architecture as described makes it possible to do so.
[0036] Since some in-hand object location systems may be limited in field of view and resolution, it can prove very beneficial to combine an in-hand object location system (75a to 75f) with a vision system (80a to 80f) as described herein below. Such an in-hand object location and vision system 70 may provide information about the 3D geometry of the object 90. This system 70 may also be used to find a different location that may provide a better grasp of the object 90. Incorporating the information of an in-hand object location system with that of a 2D/3D vision system together, facilitates a robot system to accurately and robustly pick, place and assembled objects/workpieces. This type of configuration reduces the engineering time and cost to design, build, install and tuning the system. Such a configuration may also reduce the cycle time.
[0037] In some embodiments, the plurality of in-hand tactile sensors 75a to 75f each include a layer of pressure generated illumination surfaces comprised of pressure sensitive luminescent films. Using an in-hand object location system with pressure sensitive illumination can allow easy perception of the part of an object that has been gripped without the need for an elaborate light source. Illumination surfaces may generate enough light to act as a light source for cameras 80a to 80f to receive better imagery of object 90 as it is manipulated in-hand. In some embodiments, surfaces illuminate upon coming into contact with an object 90 via a pressure-activated glow effect triggered by pressure on object 90. Tactile sensors 75a to 75f, cameras 80a to 80f and grippers 95a, 95b may be electrically and mechanically connected to a power source and control system 135 (
[0038] Referring now to
[0039] Referring now to
[0040] An offline trained model 110 (for example deep learning Convolution neural network) as shown in
[0041] In the offline training phase, the robot system automatically conducts the experiments to pick, place and assembly the parts and collects part information from 2D/3D vision system and in-hand object location and vision system 70 as well as the robot movement with the successful and fail of the picking, placing and assembly task. The initial robot picking, placing and assembly movement can come either from manual teaching or a general purpose model (the model trained for general part and tasks). In
[0042] The novel idea here is not only to use both the 2D/3D vision system and the in-hand object location system 70, which provides in hand location information after picking part, to guide robot movement 132. It allows offline training in an end-to-end model by simplifying the training phase.
[0043] Referring now to
[0044] Referring now to
[0045] In
[0046] Referring now to
[0047] The set up time step 202 is significantly shorter due to extra sensing ability available with the in-hand object location and vision system 70. Thus, the golden part training at 202 may include information from four sources working in parallel: 1) In-hand image 165, 2) vision system 170, 3) robot joint coordinates 205 and 4) synthetic information about object 210.
[0048] In
[0049] Referring now to
[0050] Referring now to
[0051] Referring now to
[0052] Using this invention, a robotic system can use a general purpose finger/gripper with or without a general purpose fixture to pick, place and assemble various parts.
[0053] The various embodiments described herein may provide the benefits of a reduction in the engineering time and cost to design, build, install and tune a special finger, or a special fixture, or a vision system for picking, placing and assembly applications in logistics, warehouse or small part assembly. Also, these embodiments may provide a reduction in cycle time since the robotic hand can detect the position of the in-hand part right after picking the part. Further, these embodiments may provide improved robustness of the system. In other words, with the highly accurate in-hand object location and geometry, the robot can adjust the placement or assembly motion to compensate for any error in the picking. Moreover, these embodiments may be easy to integrate with general purpose robot grippers, such as the robotic YUMI hand, herein incorporated by reference, for a wide range of picking, placing and assembly applications.
[0054] The techniques and systems disclosed herein may be implemented as a computer program product for use with a computer system or computerized electronic device. Such implementations may include a series of computer instructions, or logic, fixed either on a tangible/non-transitory medium, such as a computer readable medium 400 (e.g., a diskette, CD-ROM, ROM, flash memory or other memory or fixed disk) or transmittable to a computer system or a device, via a modem or other interface device, such as a communications adapter connected to a network over a medium.
[0055] The medium 300 may be either a tangible medium (e.g., optical or analog communications lines) or a medium implemented with wireless techniques (e.g., Wi-Fi, cellular, microwave, infrared or other transmission techniques). The series of computer instructions (e.g.,
[0056] Furthermore, such instructions (e.g., at 500) may be stored in any tangible memory device 505, such as semiconductor, magnetic, optical or other memory devices, and may be transmitted using any communications technology, such as optical, infrared, microwave, or other transmission technologies.
[0057] It is expected that such a computer program product may be distributed as a removable medium with accompanying printed or electronic documentation (e.g., shrink wrapped software), preloaded with a computer system (e.g., on system ROM or fixed disk), or distributed from a server or electronic bulletin board over the network (e.g., the Internet or World Wide Web). Of course, some embodiments of the invention may be implemented as a combination of both software (e.g., a computer program product) and hardware. Still other embodiments of the invention are implemented as entirely hardware, or entirely software (e.g., a computer program product).
[0058] As will be apparent to one of ordinary skill in the art from a reading of this disclosure, the present disclosure can be embodied in forms other than those specifically disclosed above. The particular embodiments described above are, therefore, to be considered as illustrative and not restrictive. Those skilled in the art will recognize, or be able to ascertain, using no more than routine experimentation, numerous equivalents to the specific embodiments described herein. Thus, it will be appreciated that the scope of the present invention is not limited to the above described embodiments, but rather is defined by the appended claims; and that these claims will encompass modifications of and improvements to what has been described.
[0059] Preferred embodiments of this invention are described herein, including the best mode known to the inventors for carrying out the invention. Variations of those preferred embodiments may become apparent to those of ordinary skill in the art upon reading the description herein. The inventors expect skilled artisans to employ such variations as appropriate, and the inventors intend for the invention to be practiced otherwise than as specifically described herein. Accordingly, this invention includes all modifications and equivalents of the subject matter recited in the claims appended hereto as permitted by applicable law. Moreover, any combination of the above-described elements in all possible variations thereof is encompassed by the invention unless otherwise indicated herein or otherwise clearly contradicted by context.