Patent classifications
G06T2207/20132
METHOD FOR EMBEDDING WATERMARK IN VIDEO DATA AND APPARATUS, METHOD FOR EXTRACTING WATERMARK IN VIDEO DATA AND APPARATUS, DEVICE, AND STORAGE MEDIUM
Disclosed in this application are a method for embedding a watermark in video data and apparatus, a method for extracting a watermark in video data and apparatus, a device, and a storage medium. The method for embedding the watermark includes: acquiring a target image frame in video data; performing time-frequency transformation on the target image frame to obtain target frequency domain data, the target frequency domain data comprising a matrix formed by frequency domain coefficients; changing the frequency domain coefficients in the target frequency domain data according to watermark data to obtain watermarked frequency domain data; performing inverse time-frequency transformation on the watermarked frequency domain data to obtain a watermarked image frame; and synthesizing watermarked video data according to the watermarked image frame.
Hand pose estimation from stereo cameras
Systems and methods herein describe using a neural network to identify a first set of joint location coordinates and a second set of joint location coordinates and identifying a three-dimensional hand pose based on both the first and second sets of joint location coordinates.
SYSTEMS AND METHODS FOR AREA-OF-INTEREST DETECTION USING SLIDE THUMBNAIL IMAGES
The subject disclosure provides systems and methods for determination of Area of Interest (AOI) for different types of input slides. Slide thumbnails may be assigned into one of five different types, and separate algorithms for AOI detection executed depending on the slide type. Slide types include ThinPrep (RTM) slides, tissue micro-array (TMA) slides, control HER2 slides with 4 cores, smear slides, and a generic slide. The slide type may be assigned based on a user input. Customized AOI detection operations are provided for each slide type. If the user enters an incorrect slide type, operations include detecting the incorrect input and executing the appropriate method. The result of each AOI detection operations provides as its output a soft-weighted image having zero intensity values at pixels that are detected as not belonging to tissue, and higher intensity values assigned to pixels detected as likely belonging to tissue regions.
Guide-assisted capture of material data
A material data collection system allows capturing of material data. For example, the material data collection system may include digital image data for materials. The material data collection system may ensure that captured digital image data is properly aligned, so that material data may be easily recalled for later use, while maintaining the proper alignment for the captured digital image. The material data collection system may include using a capture guide, to provide cues on how to orient a mobile device used with the material data collection system.
Identifying objects within images from different sources
Techniques are disclosed for providing a notification that a person is at a particular location. For example, a resident device may receive from a user device an image that shows a face of a first person, the image being captured by a first camera of the user device. The resident device may also receive, from another device having a second camera, a second image showing a portion of a face of a second person, the second camera having a viewable area showing a particular location. The resident device may determine a score indicating a level of similarity between a first set of characteristics associated with the face of the first person and a second set of characteristics associated with the face of a second person. The resident device may then provide to the user device a notification based on determining the score.
Method for Converting Landscape Video to Portrait Mobile Layout Using a Selection Interface
Described herein are systems and methods of converting media dimensions. A device may identify a set of frames from a video in a first orientation as belonging to a scene. The device may receive a selected coordinate on a frame of the set of frames for the scene. The device may identify a first region within the frame including a first feature corresponding to the selected coordinate and a second region within the frame including a second feature. The device may generate a first score for the first feature and a second score for the second feature. The first score may be greater than the second score based on the first feature corresponding to the selected coordinate. The device may crop the frame to include the first region and the second region within a predetermined display area comprising a subset of regions of the frame in a second orientation.
OBJECT IDENTIFICATION IN DIGITAL IMAGES
In an example, a digital image comprising a representation of multiple physical objects is received at a client computer. The digital image is copied into a temporary canvas. The digital image is then analyzed to identify a plurality of potential object areas, each of the potential object areas having pixels with colors similar to the other pixels within the potential object area. A minimum bounding region for each of the identified potential object areas is identified, the minimum bounding region being a smallest region of a particular shape that bounds the corresponding potential object area. The pixels within a selected minimum bounding region are cropped from the digital image. The pixels within the selected minimum bounding region are then sent to an object recognition service on a server to identify an object represented by the pixels within the selected minimum bounding region.
Method and System for Automatic Detection and Recognition of A Digital Image
An automatic measuring system containing configurable integrated circuits is able to process information via captured images. The automatic measuring system includes a metering instrument, a camera, a recognition module, and a localization module. The metering instrument has at least one display for visually displaying a number and measures the amount of measurable substance or resources (i.e., electricity and water) consumed. The camera captures an image of the number representing at least a portion the amount of measurable substance. The recognition module is operable to generate a value in response to the image and the coordinates wherein the coordinates are used to decode the image via restoring captured image to the original readout counter value. The localization module is removably or remotely coupled to the camera and operable to generate the coordinates in accordance with the image captured by the camera.
IMAGE INSPECTION DEVICE AND IMAGE INSPECTION METHOD
An image inspection device includes: an image acquisition unit to acquire an inspection target image; a geometric transformation processing unit to estimate a geometric transformation parameter for aligning a position of an inspection target in the inspection target image with a first reference image in which a position of the inspection target is known, and geometrically transform the inspection target image using the estimated geometric transformation parameter, thereby generating an aligned image in which the position of the inspection target is aligned with the first reference image; an image restoration processing unit to restore the aligned image, using an image generation network to receive an input image generated using the inspection target image and infer the aligned image as a correct image; and an abnormality determination unit to determine an abnormality of the inspection target using a difference image between the aligned image and the restored aligned image.
Recognition of activity in a video image sequence using depth information
Techniques are provided for recognition of activity in a sequence of video image frames that include depth information. A methodology embodying the techniques includes segmenting each of the received image frames into a multiple windows and generating spatio-temporal image cells from groupings of windows from a selected sub-sequence of the frames. The method also includes calculating a four dimensional (4D) optical flow vector for each of the pixels of each of the image cells and calculating a three dimensional (3D) angular representation from each of the optical flow vectors. The method further includes generating a classification feature for each of the image cells based on a histogram of the 3D angular representations of the pixels in that image cell. The classification features are then provided to a recognition classifier configured to recognize the type of activity depicted in the video sequence, based on the generated classification features.