G06F16/785

Generation of video hash
11341747 · 2022-05-24 · ·

An apparatus and method are providing for generating a hash in video in which a sample series of temporal difference are sampled in an image order. A temporal averaging is performed and a rate of change is detected to identify as distinctive events regions of high rate of change. Images having a distinctive event are labelled as distinctive images. For each image, the temporal spacing in images is calculated between that image and other distinctive images to provide a set of temporal spacings for that image; and a hash is derived for that image from that set of temporal spacings.

Methods, systems, and media for detecting a presentation of media content on a display device
11763720 · 2023-09-19 · ·

Methods, systems, and media for detecting a presentation of media content on a display device are provided. In accordance with some implementations, methods for detecting a presentation of media content on a display device are provided, the methods comprising: detecting, using a light sensor, light levels in the light sensor's surroundings; generating a signal representing the light levels; detecting, using a hardware processor, at least one variation in light levels indicative of a presentation of a video scene based on the signal; detecting at least one variation in light levels indicative of a scene change subsequent to the video scene based on the signal; and determining that media content is being presented on a display device in response to detecting the variation in light levels indicative of the presentation of the video scene and the variation in light levels indicative of the scene change.

APPARATUS FOR TRAINING RECOGNITION MODEL, APPARATUS FOR ANALYZING VIDEO, AND APPARATUS FOR PROVIDING VIDEO SEARCH SERVICE

Disclosed herein is an apparatus for analyzing a video shot. The apparatus includes at least one program, memory in which the program is recorded, and a processor for executing the program. The program may include a frame extraction unit for extracting at least one frame from a video shot, a shot composition and camera position recognition unit for predicting shot composition and a camera position for the extracted at least one frame based on a previously trained shot composition recognition model, a place and time information extraction unit for predicting a shot location and a shot time for the extracted at least one frame based on previously trained shot location recognition model and shot time recognition model, and an information combination unit for combining pieces of information, respectively predicted for the at least one frame, for each video shot and tagging the video shot with the combined pieces of information.

Digital item processing for video streams

A voice-enabled device and a display device may be utilized to determine items in video output by the display device. A frame of a video stream associated with the video may be utilized to determine a representation of a candidate object. A stream identifier and a timestamp associated with the candidate object in the video stream may be determined. The stream identifier, the timestamp, and an object identifier associated with the candidate object may be stored in a database. A first request to output the video stream via a display device may be received. A second request associated with the video stream may be received while the video stream is being output by the display device. The second request may be determined to be associated with the representation of the candidate object. The object identifier may be caused to be visually displayed via the display device.

INTERACTIVE VIDEO SURVEILLANCE AS AN EDGE SERVICE USING UNSUPERVISED FEATURE QUERIES
20220292827 · 2022-09-15 ·

A method for querying data obtained from a distributed sensor network, comprising: receiving sensor data representing an aspect of an environment with a sensor of the distributed sensor network; communicating a representation of the sensor data to a fog node through an automated communication network; determining, by the fog node, a correspondence of a query received through the automated communication network to characteristics of the representation of the sensor data; and selectively communicating, in response to the query, at least one of: the sensor data having the determined characteristics corresponding to the query, an identification of the sensor data having the determined characteristics corresponding to the query, and the data representing the sensor data having the determined characteristics corresponding to the query.

VIDEO DATA SEARCH USING COLOR WHEEL ASSOCIATIONS
20220083594 · 2022-03-17 ·

Systems, methods, and software described herein to manage preferences for video search. In one implementation, a video processing service may receive a request for video data that satisfies one or more criteria, wherein the one or more criteria includes at least one color for an object. In response to the request, the video processing service may identify one or more secondary colors associated with the at least one color and update the one or more criteria to include the one or more secondary colors. The video processing service may then select the video data from a video source that satisfies the updated one or more criteria.

SMART IP CAMERA WITH COLOR NIGHT MODE
20220094847 · 2022-03-24 ·

An apparatus includes a camera and a processor circuit. The camera may be configured to capture color images in response to visible light and monochrome infrared images in response to infrared light. The processor circuit may be configured to extract color features from the color images and add color to corresponding monochrome features detected in the monochrome infrared images.

SUMMARIZING VIDEO CONTENT

Systems and methods of automatically extracting summaries of video content are described herein. A data processing system can access, from a video database, a first video content element including a first plurality of frame. The data processing system can select an intervallic subset of the first plurality of frames of the first video content element. The data processing system can calculate, for each of a plurality of further subsets comprising a predetermined number of frames from the intervallic subset, a score for the further subset. The data processing system can identify, from the plurality of further subsets, a further subset having a highest score. The data processing system can select a portion of the first video content element comprising the frames of the further subset having the highest score. The data processing system can generate a second video content element comprising the selected portion of the first video content element.

Metric-based recognition, systems and methods
11062169 · 2021-07-13 · ·

Apparatus, methods and systems of object recognition are disclosed. Embodiments of the inventive subject matter generates map-altered image data according to an object-specific metric map, derives a metric-based descriptor set by executing an image analysis algorithm on the map-altered image data, and retrieves digital content associated with a target object as a function of the metric-based descriptor set.

Summarizing video content

Systems and methods of automatically extracting summaries of video content are described herein. A data processing system can access, from a video database, a first video content element including a first plurality of frame. The data processing system can select an intervallic subset of the first plurality of frames of the first video content element. The data processing system can calculate, for each of a plurality of further subsets comprising a predetermined number of frames from the intervallic subset, a score for the further subset. The data processing system can identify, from the plurality of further subsets, a further subset having a highest score. The data processing system can select a portion of the first video content element comprising the frames of the further subset having the highest score. The data processing system can generate a second video content element comprising the selected portion of the first video content element.