G11B27/036

METHOD AND SYSTEM FOR AUTOMATIC PRE-RECORDATION VIDEO REDACTION OF OBJECTS
20230046913 · 2023-02-16 · ·

A system and a method for automatic video redaction are provided herein. The method may include: receiving, an input video comprising a sequence of frames captured by a camera, wherein the input video includes live video obtained directly from the camera, wherein recordation of the video directly from the camera is disabled; performing visual analysis of the input video, to detect portions of the frames of the input video in which one of a plurality of predefined objects or a descriptor thereof is detected; generating a redacted input video by replacing the portions of the frames with new portions of another visual content; and recording the redacted input video on a data storage device, wherein the generating of thethe redacted input video, is carried out by a computer processor, after the input video is captured by the camera and before the recording of the redacted input video on the data storage device.

Facial synchronization utilizing deferred neural rendering

Techniques are disclosed for performing video synthesis of audiovisual content. In an example, a computing system may determine first facial parameters of a face of a particular person from a first frame in a video shot, whereby the video shot shows the particular person speaking a message. The system may determine second facial parameters based on an audio file that corresponds to the message being spoken in a different way from the video shot. The system may generate third facial parameters by merging the first and the second facial parameters. The system may identify a region of the face that is associated with a difference between the first and second facial parameters, render the region of the face based on a neural texture of the video shot, and then output a new frame showing the face of the particular person speaking the message in the different way.

Facial synchronization utilizing deferred neural rendering

Techniques are disclosed for performing video synthesis of audiovisual content. In an example, a computing system may determine first facial parameters of a face of a particular person from a first frame in a video shot, whereby the video shot shows the particular person speaking a message. The system may determine second facial parameters based on an audio file that corresponds to the message being spoken in a different way from the video shot. The system may generate third facial parameters by merging the first and the second facial parameters. The system may identify a region of the face that is associated with a difference between the first and second facial parameters, render the region of the face based on a neural texture of the video shot, and then output a new frame showing the face of the particular person speaking the message in the different way.

Image processing apparatus, image processing method and medium

An object of one embodiment of the present disclosure is to provide a product with a high added value to a user by preventing an unnatural character string from being combined, combination of no character string, and the like in a case where there is no voice or almost no voice before or after an image selected from within a moving image. One embodiment of the present disclosure is an image processing apparatus including: a selection unit configured to select, from a moving image including a plurality of frames, a part of the moving image; an extraction unit configured to extract a voice during a predetermined time corresponding to the selected part in the moving image; and a combination unit configured to combine a character string based on a voice extracted by the extraction unit, with the part of the moving image selected by the selection unit.

Image processing apparatus, image processing method and medium

An object of one embodiment of the present disclosure is to provide a product with a high added value to a user by preventing an unnatural character string from being combined, combination of no character string, and the like in a case where there is no voice or almost no voice before or after an image selected from within a moving image. One embodiment of the present disclosure is an image processing apparatus including: a selection unit configured to select, from a moving image including a plurality of frames, a part of the moving image; an extraction unit configured to extract a voice during a predetermined time corresponding to the selected part in the moving image; and a combination unit configured to combine a character string based on a voice extracted by the extraction unit, with the part of the moving image selected by the selection unit.

Dynamically modeling an object in an environment from different perspectives
11557080 · 2023-01-17 · ·

An object can be simulated in an environment using a three-dimensional model of the object as viewed from a virtual camera at a position in the environment. The position in the environment can be determined using user input or through visual analysis of a video recording. Composite frames depicting the modeled object may be played back based on the orientation of the playback device.

Dynamically modeling an object in an environment from different perspectives
11557080 · 2023-01-17 · ·

An object can be simulated in an environment using a three-dimensional model of the object as viewed from a virtual camera at a position in the environment. The position in the environment can be determined using user input or through visual analysis of a video recording. Composite frames depicting the modeled object may be played back based on the orientation of the playback device.

Systems and methods for automatic mixing of media

A first device includes one or more processors and memory storing one or more programs configured to be executed by the one or more processors. The one or more programs include instructions for receiving, from a second device, audio mix information for a first audio item and receiving, from the second device, an indication that the first audio item is to be mixed with a second audio item distinct from the first audio item. In response to the indication, the one or more programs include instructions for transmitting to the second device an audio stream including the first audio item and the second audio item mixed in accordance with the audio mix information.

Systems and methods for automatic mixing of media

A first device includes one or more processors and memory storing one or more programs configured to be executed by the one or more processors. The one or more programs include instructions for receiving, from a second device, audio mix information for a first audio item and receiving, from the second device, an indication that the first audio item is to be mixed with a second audio item distinct from the first audio item. In response to the indication, the one or more programs include instructions for transmitting to the second device an audio stream including the first audio item and the second audio item mixed in accordance with the audio mix information.

Generating a stitched data stream
11558678 · 2023-01-17 · ·

Systems and methods provide for receiving a request for an alternate data stream of a plurality of individual data streams than an active data stream currently being displayed on a computing device, during display on the computing device of a stitched data steam comprising the plurality of individual data streams associated with a common audio timeline. The systems and methods further providing for determining a subset of the plurality of individual data streams of the stitched data stream associated with a time period of the active data stream in the common audio timeline, selecting the alternate data stream from the subset of the plurality of individual data streams, and providing the alternate data stream to the computing device, wherein the display of the active data stream on the computing device transitions to the alternate data stream on the computing device in the common audio timeline.