Patent classifications
G06V10/98
INFORMATION PROCESSING DEVICE AND INFORMATION PROCESSING METHOD
An information processing device according to the present disclosure includes: an acquisition unit that acquires outline information indicating an outline of a user who makes a body motion; and a specification unit that specifies, among body parts, a main part corresponding to the body motion and a related part, which is to be a target of correction processing of motion information corresponding to the body motion, on the basis of the outline information acquired by the acquisition unit.
VIDEO PROCESSING APPARATUS, METHOD AND COMPUTER PROGRAM
A video processing apparatus configured to process a stream of video surveillance data, wherein the video surveillance data includes metadata associated with video data, the metadata describing at least one object in the video data. The apparatus comprises means for applying an image assessment algorithm to generate a reliability score for the metadata, and associating the reliability score with the metadata. The image assessment algorithm generates the reliability score based on an assessment of the image quality of the video data to which the metadata relates to indicate a likelihood that the metadata accurately describes the object. An image enhancement module applies image enhancement to video data if the reliability score of metadata associated with the video data indicates a low likelihood that the metadata accurately describes the object.
OBJECT DETECTION METHOD AND OBJECT DETECTION DEVICE
A detection method of detecting a defined object from an image, includes estimating, on the image, an extreme point area including a boundary feature point that satisfies a criterion related to a boundary of the object.
OBJECT DETECTION METHOD AND OBJECT DETECTION DEVICE
A detection method of detecting a defined object from an image, includes estimating, on the image, an extreme point area including a boundary feature point that satisfies a criterion related to a boundary of the object.
SYSTEM AND METHOD FOR DETECTING ERRORS IN A TASK WORKFLOW FROM A VIDEO STREAM
A system for detecting errors in task workflows from a real time video feed records. The video feed that shows a plurality of steps being performed to accomplish a plurality of tasks through an automation process system. The system splits the video feed into a plurality of video recordings which are valid breakpoints determined through cognitive Machine Learning Engine, where each video recording shows a single task. For each task from among the plurality of tasks, the system determines whether the task fails and the exact point of failure for that task. If the system determines that the task fails, the system determines a particular step where the task fails. The system flags the particular step as a failed step. The system reports the flagged step for troubleshooting.
Method and system for distributed learning and adaptation in autonomous driving vehicles
The present teaching relates to system, method, medium for in-situ perception in an autonomous driving vehicle. A plurality of types of sensor data acquired continuously by a plurality of types of sensors deployed on the vehicle are first received, where the plurality of types of sensor data provide information about surrounding of the vehicle. Based on at least one model, one or more items are tracked from a first of the plurality of types of sensor data acquired by one or more of a first type of the plurality of types of sensors, wherein the one or more items appear in the surrounding of the vehicle. At least some of the one or more items are then automatically labeled on-the-fly via either cross modality validation or cross temporal validation of the one or more items and are used to locally adapt, on-the-fly, the at least one model in the vehicle.
System and method for detecting errors in a task workflow from a video stream
A system for detecting errors in task workflows from a real time video feed records. The video feed that shows a plurality of steps being performed to accomplish a plurality of tasks through an automation process system. The system splits the video feed into a plurality of video recordings which are valid breakpoints determined through cognitive Machine Learning Engine, where each video recording shows a single task. For each task from among the plurality of tasks, the system determines whether the task fails and the exact point of failure for that task. If the system determines that the task fails, the system determines a particular step where the task fails. The system flags the particular step as a failed step. The system reports the flagged step for troubleshooting.
METHODS FOR MOBILE IMAGE CAPTURE OF VEHICLE IDENTIFICATION NUMBERS IN A NON-DOCUMENT
Various embodiments disclosed herein are directed to methods of capturing Vehicle Identification Numbers (VIN) from images captured by a mobile device. Capturing VIN data can be useful in several applications, for example, insurance data capture applications. There are at least two types of images supported by this technology: (1) images of documents and (2) images of non-documents.
Information processing apparatus, information processing method, and program
An information processing apparatus (100) includes an acquisition unit (122) that acquires a first image from which person region feature information regarding a region including other than a face of a retrieval target person is extracted, a second image in which a collation result with the person region feature information indicates a match, and a facial region is detected, and result information indicating a collation result between face information stored in a storage unit and face information extracted from the facial region, and a display processing unit (130) that displays at least two of the first image, the second image, and the result information on an identical screen.
Information processing apparatus, information processing method, and program
An information processing apparatus (100) includes an acquisition unit (122) that acquires a first image from which person region feature information regarding a region including other than a face of a retrieval target person is extracted, a second image in which a collation result with the person region feature information indicates a match, and a facial region is detected, and result information indicating a collation result between face information stored in a storage unit and face information extracted from the facial region, and a display processing unit (130) that displays at least two of the first image, the second image, and the result information on an identical screen.