H04N9/8211

Information processing apparatus and information processing method to modify an image based on audio data

An information processing apparatus is provided that includes a playback unit to play back music data, an analysis unit to analyze a feature of a relevant image of the music data, an image correction unit to perform image correction with use of any of a plurality of correction types, a storage unit to store one or more than one image, a selection unit to select a correction type corresponding to the feature of the relevant image analyzed by the analysis unit from the plurality of correction types, a correction control unit to cause the image correction unit to perform image correction of an image stored in the storage unit with use of the correction type selected by the selection unit, and an output unit to output the image corrected by the image correction unit.

Reusable envelope
09840352 · 2017-12-12 ·

A reusable envelope has one or more spaces thereon where address, return labels and/or postage can be removably placed or permanently affixed such as for finite number of shipments using a permanently affixed postage area, for example. The reusable envelope also includes a resealable closure allowing the interior of the envelope to be closed and reopened multiple times for multiple uses (shipments, receiving) or optionally to use a finite number of times by exposing the permanent adhesive to the flat and mating surface depending on user preference.

Controlling sounds of individual objects in a video

A method for modifying a sound produced by a sound source in a video includes capturing video and audio of a scene is disclosed. Audio is captured using a microphone array. A sound source is isolated and a direction of arrival of the sound source with respect to a capture location is identified. One or more visual objects in the captured video are identified. One of the isolated sound sources is associated with one of the identified visual objects. An input identifying one of the isolated sound sources is received during playing of the captured video and audio. The input includes a command. Responsive to receiving the input, an attribute of the identified isolated sound source is modified. The input may identify a visual object associated with a sound source. A system and article of manufacture are also disclosed.

Generation of correlated keyword and image data
09837125 · 2017-12-05 · ·

A data processing apparatus includes a text acquisition mechanism acquiring texts to be used as keywords which will be subject to audio retrieval, the texts being related to contents corresponding to contents data including image data and audio data; a keyword acquisition mechanism acquiring the keywords from the texts; an audio retrieval mechanism retrieving utterance of the keywords from the audio data of the contents data and acquiring timing information representing the timing of the utterance of the keywords of which the utterance is retrieved; and a playback control mechanism generating, from image data around the time represented by the timing information, representation image data of a representation image which will be displayed together with the keywords and performing playback control of displaying the representation image corresponding to the representation image data together with the keywords which are uttered at the time represented by the timing information.

Billiard table lighting and game play monitor

A billiard table top lighting apparatus provides substantially uniform lighting across the surface of a billiard table surface. The frame may support one or more cameras, one or more motion sensors, one or more microphones, and/or one or more computing devices to enable any of a variety of innovative features. Such features could include automatic game play recording from one or more perspectives, merged video track storage for replay, review, and analysis, automatic lighting and dimming control, control of the apparatus from any mobile device, automatic provision of a shot clock, and the like.

RECORDING DEVICE AND MULTIPLEXING METHOD
20170332039 · 2017-11-16 ·

A recording device includes a video encoding unit configured to compress and then encode video data pieces by a first unit of data, an audio encoding unit configured to compress and then encode audio data pieces in a second unit of data, and a multiplexing unit. The multiplexing unit selects the audio data pieces by the second unit of data from the audio data pieces in such a way that an amount of temporal shift between the respective video data pieces and the audio data pieces corresponding to the respective video data pieces will become less than or equal to a time length of the second unit of data and then multiplexes the selected audio data pieces with the video data.

VIDEO GENERATION METHOD, APPARATUS AND TERMINAL
20170332020 · 2017-11-16 ·

Embodiments of the present disclosure provide a video generation method, device and terminal. Embodiments of the present disclosure may perform effect processing for the image frames obtained from a video during shooting of the video or after the shooting of the video is finished using the selected video effect template, and compose the processed image frames to obtain the composite video. The problem of only presenting contents shot using a camera during the recording of a song and being not able to provide customized functions satisfying users' requirements may be solved. The effect of performing effect processing for all image frames or partial image frames in the shot video using the video effect template selected by the user to obtain the composite video satisfying the users' requirements may be achieved.

SURROUND VIDEO PLAYBACK
20170318274 · 2017-11-02 ·

Methods and systems are disclosed including a computing device configured to allow a user to view a multi-stream video from a selected angle/direction with respect to the contents of the multi-stream video, under the user's control. The multi-stream video is generated using multiple Image Acquisition Devices (IAD), such as cameras, simultaneously, consecutively, or independently filming a scene, each IAD having a different position with respect to each of the other IADs. Each image data stream obtained from each IAD may be uniquely identified to allow selective real-time playback of image data streams under user control. Each image data stream represents a corresponding viewing angle to the user. The user may dynamically change the selection of an image stream, and thus the viewing angle, while viewing a recorded scene. Multiple image streams of the same scene may be selected and viewed simultaneously to provide 3D or other visual effects.

Perspective view entertainment system and method
09805767 · 2017-10-31 ·

In a method according to the present disclosure, a director's view version of a film is recorded. Then the film is recorded from the viewpoint of a different characters in the film. The director's view version and character view versions are time synched to create a film that allows a user to switch at any time between the director's view version and one or more of the character view versions during viewing of the film. A system according to the method uses a director's view camera and at least one character view camera to record a scene. A recording processor communicates with the cameras and receives and stores in memory director's view camera data and character view camera data. The recording processor further time-synchs the director's view camera data and character view camera data. A viewing system has a viewing screen, a viewer-operated controller, and a viewing processor configured to display on the viewing screen at least one of the perspective views of the film scene and to switch between the perspective views of the film scene upon actuation by the viewer of the viewer-operated controller.

Systems and methods for adjusting dubbed speech based on context of a scene

Systems and methods are disclosed herein for detecting dubbed speech in a media asset and receiving metadata corresponding to the media asset. The systems and methods may determine a plurality of scenes in the media asset based on the metadata, retrieve a portion of the dubbed speech corresponding to the first scene, and process the retrieved portion of the dubbed speech corresponding to the first scene to identify a speech characteristic of a character featured in the first scene. Further, the systems and methods may determine whether the speech characteristic of the character featured in the first scene matches the context of the first scene, and if the match fails, perform a function to adjust the portion of the dubbed speech so that the speech characteristic of the character featured in the first scene snatches the context of the first scene.