Patent classifications
G06V10/24
INTERACTIVE VIDEO PLAYBACK TECHNIQUES TO ENABLE HIGH FIDELITY MAGNIFICATION
Responsive to a zoom command when presenting a first video, a second video is combined with the first video and presented. The first and second videos are generated from substantially the same camera location as each other at substantially the same time with substantially the same resolution. However, the second video is generated by a physical or virtual lens having a field of view (FOV) smaller than the FOV of a physical or virtual lens used in generating the first video. Modules are described for using alignment metrics to correctly place the second video over the inner video and make it appear seamless.
Method and system for facilitating access to recorded data
The present invention relates to a method and system for facilitating access to recorded data. The system comprises an interface and a processing device. The interface is arranged to receive data and the processing device is arranged to separate the received data in data subsets, compress each data subset and assign an identifier to each compressed data subset, thereby creating data units each comprising a compressed data subset and an associated identifier, the processing device further being arranged to establish an index on the basis of the assigned identifiers.
Method and system for facilitating access to recorded data
The present invention relates to a method and system for facilitating access to recorded data. The system comprises an interface and a processing device. The interface is arranged to receive data and the processing device is arranged to separate the received data in data subsets, compress each data subset and assign an identifier to each compressed data subset, thereby creating data units each comprising a compressed data subset and an associated identifier, the processing device further being arranged to establish an index on the basis of the assigned identifiers.
AUTHENTICATION METHOD, NON-TRANSITORY COMPUTER-READABLE STORAGE MEDIUM FOR STORING AUTHENTICATION PROGRAM, AND AUTHENTICATION APPARATUS
An authentication method executed by a computer, the authentication method including: extracting, when a captured image of a living body is acquired, a biometric image included in a region that corresponds to the living body from the captured image; and performing authentication of the living body on the basis of the extracted biometric image and a position of the biometric image in the captured image.
PERCEPTION SYSTEM WITH ATTENTION MODULE FOR PROCESSING VISUAL DATA
A perception system is adapted to receive visual data from a camera and includes a controller having a processor and tangible, non-transitory memory on which instructions are recorded. A subsampling module, an object detection module and an attention module are each selectively executable by the controller. The controller is configured to sample an input image from the visual data to generate a rescaled whole image frame, via the subsampling module. The controller is configured to extract feature data from the rescaled whole image frame, via the object detection module. A region of interest in the rescaled whole image frame is identified, based on an output of the attention module. The controller is configured to generate a first image based on the rescaled whole image frame and a second image based on the region of interest, the second image having a higher resolution than the first image.
Combine Orientation Tracking Techniques of Different Data Rates to Generate Inputs to a Computing System
A system to combine inertial-based measurements and optical-based measurements via a Kalman-type filter. For example, a sensor module uses an inertial measurement unit to generate first positions and first orientations of the sensor module at a first time interval during a first period of time containing multiple of the first time interval. At least one camera is used to capture images of the sensor module at a second time interval, larger than the first time interval, during the first period of time containing multiple of the second interval. Second positions and second orientations of the sensor module during the first period of time are computed from the images. The filter receives the first positions, the first orientations, the second positions, and the second orientations to generate estimates of position and orientation of the sensor module at a time interval no smaller than the first time interval.
Multi-Angle Object Recognition
Methods, systems, and apparatus for controlling smart devices are described. In one aspect a method includes capturing, by a camera on a user device, a plurality of successive images for display in an application environment of an application executing on the user device, performing an object recognition process on the images, the object recognition process including determining that a plurality of images, each depicting a particular object, are required to perform object recognition on the particular object, and in response to the determination, generating a user interface element that indicates a camera operation to be performed, the camera option capturing two or more images, determining that a user, in response to the user interface element, has caused the indicated camera operation to be performed to capture the two or more images, and in response, determining whether a particular object is positively identified from the plurality of images.
SELECTIVELY INCREASING DEPTH-OF-FIELD IN SCENES WITH MULTIPLE REGIONS OF INTEREST
The present disclosure provides systems, apparatus, methods, and computer-readable media that support multi-frame depth-of-field (MF-DOF) for deblurring background regions of interest (ROIs), such as background faces, that may be blurred due to a large aperture size or other characteristics of the camera used to capture the image frame. The processing may include the use of two image frames obtained at two different focus points corresponding to the multiple ROIs in the image frame. The corrected image frame may be determined by deblurring one or more ROIs of the first image frame using an AI-based model and/or local gradient information. The MF-DOF may allow selectively increasing a depth-of-field (DOF) of an image to provide focused capture of multiple regions of interest, without causing a reduction in aperture (and subsequently an amount of light available for photography) or background blur that may be desired for photography.
SELECTIVELY INCREASING DEPTH-OF-FIELD IN SCENES WITH MULTIPLE REGIONS OF INTEREST
The present disclosure provides systems, apparatus, methods, and computer-readable media that support multi-frame depth-of-field (MF-DOF) for deblurring background regions of interest (ROIs), such as background faces, that may be blurred due to a large aperture size or other characteristics of the camera used to capture the image frame. The processing may include the use of two image frames obtained at two different focus points corresponding to the multiple ROIs in the image frame. The corrected image frame may be determined by deblurring one or more ROIs of the first image frame using an AI-based model and/or local gradient information. The MF-DOF may allow selectively increasing a depth-of-field (DOF) of an image to provide focused capture of multiple regions of interest, without causing a reduction in aperture (and subsequently an amount of light available for photography) or background blur that may be desired for photography.
COMPUTER-IMPLEMENTED METHOD, COMPUTER PROGRAM PRODUCT AND SYSTEM FOR ANALYZING VIDEOS CAPTURED WITH MICROSCOPIC IMAGING
A computer-implemented method is provided for analyzing videos of a living system captured with microscopic imaging. The method can include obtaining a base dataset including one or more videos captured with microscopic imaging with at least one of the one or more videos including a cellular event, and cropping out, from the base dataset, sub-videos including one or more objects of interest that may be involved in the cellular event. An artificial neural network (ANN) model can be trained using the plurality of selected sub-videos as training data, to perform unsupervised video alignment, a query sub-video can be aligned using the trained ANN model, and a determination can be made whether or not the query sub-video includes the cellular event.