Patent classifications
G06V20/647
Object recognition processing apparatus and method, and object picking apparatus and method
An object recognition processing apparatus includes: a model data acquisition unit configured to acquire three-dimensional model data of an object; a measurement unit configured to acquire measurement data including three-dimensional position information of the object; a position/orientation recognition unit configured to recognize a position/orientation of the object based on the three-dimensional model data and the measurement data; a similarity score calculation unit configured to calculate a similarity score indicating a degree of similarity between the three-dimensional model data and the measurement data in a position/orientation recognition result of the object; a reliability calculation unit configured to calculate an index indicating a feature of a three-dimensional shape of the object, and calculate a reliability of the similarity score based on the index; and an integrated score calculation unit configured to calculate an integrated score indicating a quality of the position/orientation recognition result based on the similarity score and the reliability.
Scanning of 3D objects with a second screen device for insertion into a virtual environment
A method for inserting a virtual object into a virtual environment is provided, including: generating a scan of a real world object; processing the scan to identify a type of the real world object; based on the identified type of the real world object, generating a virtual object that resembles the real world object; based on the identified type of the real world object, assigning a functionality to the virtual object, the functionality defining an action capable of being performed by the virtual object in a virtual environment; deploying the virtual object in the virtual environment.
FACE SWAPPING WITH NEURAL NETWORK-BASED GEOMETRY REFINING
Various embodiments set forth systems and techniques for changing a face within an image. The techniques include receiving a first image including a face associated with a first facial identity; generating, via a machine learning model, at least a first texture map and a first position map based on the first image; rendering a second image including a face associated with a second facial identity based on the first texture map and the first position map, wherein the second facial identity is different from the first facial identity.
Information processing device, information processing method, and information processing program
An information processing device (100) according to the present disclosure includes: an acquisition unit (141) configured to acquire a first image including a content image of an ear of a user; and a calculation unit (142) configured to calculate, based on the first image acquired by the acquisition unit (141), a head-related transfer function corresponding to the user by using a learned model having learned to output a head-related transfer function corresponding to an ear when an image including a content image of the ear is input.
Robotic system architecture and control processes
A system includes a first sensor having a fixed location relative to a workspace, a second sensor, at least one robotic manipulator coupled to a manipulation tool, and a control system in communication with the at least one robotic manipulator. The control system is configured to determine a location of a workpiece in the workspace based on first sensor data from the first sensor and a three-dimensional (3D) model corresponding to the workpiece. The control system is configured to map a set of 2D coordinates from a second 2D image from the second sensor to a set of 3D coordinates based on the location, and to generate one or more control signals for the at least one robotic manipulator based on the set of 3D coordinates.
Skeletal tracking using previous frames
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and a method for detecting a pose of a user. The program and method include operations comprising receiving a monocular image that includes a depiction of a body of a user; detecting a plurality of skeletal joints of the body based on the monocular image; accessing a video feed comprising a plurality of monocular images received prior to the monocular image; filtering, using the video feed, the plurality of skeletal joints of the body detected based on the monocular image; and determining a pose represented by the body depicted in the monocular image based on the filtered plurality of skeletal joints of the body.
User data validation for digital identifications
In general, one innovative aspect of the subject matter described in this specification may be embodied in methods that may include validating user data pages extracted from a digital identification in circumstances where a user device that includes the digital identification is either unavailable or presently lacks network connectivity. For instance, an authorized device may be used to extract user data pages from the digital identification by either exchanging communications with the user device using a proximity-based data exchange protocol, or by using a physical identification card to identify the digital identification on a user record. The user data pages may then be validated by comparing checksums associated with user data pages against the checksums within the user record, and decrypting the user data pages using a decryption key that is variably designated by a security status assigned to the digital identification.
Inverse path tracing for material and lighting estimation
In one embodiment, a computing system accesses a three-dimensional (3D) model of an environment, the 3D model comprising a virtual representation of an object in the environment. The computing system accesses an image of the object captured by a camera from a camera pose. The computing system accesses light source parameters associated with a virtual representation of a light source in the environment. The computing system renders, using the 3D model, pixels associated with the virtual representation of the object based on the light source parameters, the pixels being rendered from a virtual perspective corresponding to the camera pose. The computing system determines updated light source parameters based on a comparison of the rendered pixels to corresponding pixels located in the image of the object.
Extended reality space generating apparatus and method
An extended reality space generating apparatus and method are provided. The extended reality space generating apparatus generates a plurality of plane plates, a plate coordinate and a normal vector corresponding to each of the plane plates based on a plurality of point clouds, wherein the point clouds correspond to a real space. The extended reality space generating apparatus compares the plate coordinates and the normal vectors of the plane plates in a visual window to generate an effective plane plate set. The extended reality space generating apparatus generates an extended reality space corresponding to the real space based on the effective plane plate set.
3D BOUNDING BOX RECONSTRUCTION METHOD, 3D BOUNDING BOX RECONSTRUCTION SYSTEM AND NON-TRANSITORY COMPUTER READABLE MEDIUM
A 3D bounding box reconstruction method includes obtaining masks corresponding to a target object in images, obtaining a trajectory direction of the target object according to the masks, generating a target contour according to one of the masks, transforming the target contour into a transformed contour using a transformation matrix, obtaining a first bounding box according to the transformed contour and the trajectory direction, transforming the first bounding box into a second bounding box corresponding to the target contour using the transformation matrix, obtaining first reference points according to the target contour and the second bounding box, transforming the first reference points into second reference points using the transformation matrix, obtaining a third bounding box using the second reference points, transforming the third bounding box into a fourth bounding box using the transformation matrix, and obtaining a 3D bounding box using the second bounding box and the fourth bounding box.