H04N9/74

System and Method for Automated Video Editing
20170236549 · 2017-08-17 · ·

A system and method for automated video editing. A reference media is selected and analyzed. At least one video may be acquired, and thereby synced to the reference audio media. Once synced, audio analysis is used to assemble an edited video. The audio analysis can include information, including user inputs, video analysis, and metadata. The system and method for automated video editing may be applied to collaborative creation, simulated stop motion animation, and real-time implementations.

System and Method for Automated Video Editing
20170236549 · 2017-08-17 · ·

A system and method for automated video editing. A reference media is selected and analyzed. At least one video may be acquired, and thereby synced to the reference audio media. Once synced, audio analysis is used to assemble an edited video. The audio analysis can include information, including user inputs, video analysis, and metadata. The system and method for automated video editing may be applied to collaborative creation, simulated stop motion animation, and real-time implementations.

Replacing moving objects with background information in a video scene
11430132 · 2022-08-30 · ·

Methods, an apparatus, and software media are provided for removing unwanted information such as moving or temporary foreground objects from a video sequence. The method performs, for each pixel, a statistical analysis to create a background data model whose color values can be used to detect and remove the unwanted information. This includes determining a prevalent color cluster from among k clusters of color values for the pixel in successive frames. The method uses k-means clustering. To replace the unwanted information, the method iterates frames to find frames in which a pixel's color value is not included in the prevalent color cluster. In those frames, it replaces the pixel's color value with a value from the prevalent color cluster.

Method and system for multiple f-number lens

An imaging lens includes one or more lens elements configured to receive and focus light in a first wavelength range reflected off of one or more first objects onto an image plane, and to receive and focus light in a second wavelength range reflected off of one or more second objects onto the image plane. The imaging lens further includes an aperture stop and a filter positioned at the aperture stop. The filter includes a central region and an outer region surrounding the central region. The central region of the filter is characterized by a first transmission band in the first wavelength range and a second transmission band in the second wavelength range. The outer region of the filter is characterized by a third transmission band in the first wavelength range and substantially low transmittance values in the second wavelength range.

Systems and methods for implementing augmented reality
09721387 · 2017-08-01 · ·

A method implemented in a video playback system is described for incorporating augmented reality (AR) into a video stream. The method comprises determining a target pattern, determining an inner pattern in the target pattern, determining a relationship between the target pattern and the inner pattern, and receiving, by the video playback system, frames of the video stream. For each frame within the frame sequence, binarization is performed according to a predetermined threshold. Based on whether a location of the target pattern can be determined, a location of the inner pattern is determined. Based on the location of the inner pattern on received frames and the determined relationship between the target pattern and the inner pattern, a location of the target pattern is determined. The method further comprises displaying a virtual object with the target pattern on an output device based on the location of the target pattern.

Method and device for shooting image, and storage medium

A method for shooting image, applied to an electronic device in which an image capturing device is mounted, includes: receiving a first input from a user; responsive to the first input, acquiring a first image and a second image captured by the image capturing device for a target object, a framing range of the first image being same as a framing range of the second image; and performing synthesis processing on the first image and the second image to generate a target image, the target image including a first object image of the target object in the first image and a second object image of the target object in the second image.

LIGHTING, COLOR VECTOR, AND VIRTUAL BACKGROUND CORRECTION DURING A VIDEO CONFERENCE SESSION
20220232189 · 2022-07-21 · ·

An information handling system executing a multimedia multi-user collaboration application (MMCA) may include a memory; a power management unit; a camera to capture video of a user participating in a video conference session; a processor configured to execute code instructions of a trained intelligent collaboration contextual session management system (ICCSMS) neural network to receive as inputs: the type of AV processing instruction modules enabled descriptive of how to visually transform a video frame during a video conference session executed by a multimedia multi-user collaboration application; and sensor data from a plurality of sensors including an ambient light sensor to detect ambient light around a participant of the video conference session and a color senser to detect color vectors in the video frame; the processor applies AV processing instruction adjustments to the enabled AV processing instruction modules received as output from the trained ICC SMS machine learning module to adjust the lighting and color vectors of the video frame based on the sensor inputs and the type of AV processing instruction modules.

Targeted video streaming post-production effects

Visual effects such as bleach bypass, sepia tone conversion, cross processing, custom effects, and many others may be applied to video streams before they are displayed to viewers. In some examples, different visual effects may be applied to the same underlying video content, such as on a viewer-by-viewer basis or at other levels of granularity, thereby allowing certain particular visual effects to be targeted to particular viewers/devices. The visual effects may be applied by one or more of a content provider, a video streaming service, one or more viewers, or by other entities. The visual effects may be applied based at least in part on instructions from the provider, such as one or more tags, for example that may be issued via an interface provided by a video streaming service.

PLATE RECONSTRUCTION OF OBSCURED VIEWS OF A MAIN IMAGING DEVICE USING CAPTURE DEVICE INPUTS OF THE SAME SCENE

An imagery processing system obtains capture inputs from capture devices that might have capture parameters and characteristics that differ from those of a main imagery capture device. By normalizing outputs of those capture devices, potentially arbitrary capture devices could be used for reconstructing portions of a scene captured by the main imagery capture device when reconstructing a plate of the scene to replace an object in the scene with what the object obscured in the scene. Reconstruction could be of one main image, a stereo pair of images, or some number, N, of images where N>2.

PLATE RECONSTRUCTION OF OBSCURED VIEWS OF A MAIN IMAGING DEVICE USING CAPTURE DEVICE INPUTS OF THE SAME SCENE

An imagery processing system obtains capture inputs from capture devices that might have capture parameters and characteristics that differ from those of a main imagery capture device. By normalizing outputs of those capture devices, potentially arbitrary capture devices could be used for reconstructing portions of a scene captured by the main imagery capture device when reconstructing a plate of the scene to replace an object in the scene with what the object obscured in the scene. Reconstruction could be of one main image, a stereo pair of images, or some number, N, of images where N>2.