Patent classifications
H04N9/8715
Image processing apparatus, image processing method and medium
An object of one embodiment of the present disclosure is to provide a product with a high added value to a user by preventing an unnatural character string from being combined, combination of no character string, and the like in a case where there is no voice or almost no voice before or after an image selected from within a moving image. One embodiment of the present disclosure is an image processing apparatus including: a selection unit configured to select, from a moving image including a plurality of frames, a part of the moving image; an extraction unit configured to extract a voice during a predetermined time corresponding to the selected part in the moving image; and a combination unit configured to combine a character string based on a voice extracted by the extraction unit, with the part of the moving image selected by the selection unit.
Selection of a prerecorded media file for superimposing into a video
In a method for selecting of a prerecorded media file for superimposing into a video, a video of a scene is displayed on a display device of a mobile electronic device. A location of the scene is determined. A prerecorded video file is selected based at least in part on the location. The prerecorded video file is superimposed over the video, such that the video is partially obscured by the prerecorded video file. The prerecorded video file is played while displaying the video, such that the prerecorded video file and a non-obscured portion of the video are rendered simultaneously.
COMPUTING DEVICE AND CORRESPONDING METHOD FOR GENERATING DATA REPRESENTING TEXT
An example method involves (i) accessing first data representing text, wherein the text defines at least one position representing a particular type of grammatical break between two portions of the text; (ii) identifying, from among the at least one position, a position that is closest to a target position within the text; (iii) based on the identified position within the text, generating second data that represents a proper subset of the text, wherein the proper subset extends from an initial position within the text to the identified position within the text; and (iv) providing output based on the generated second data.
Systems and Methods for Previewing Newly Captured Image Content and Reviewing Previously Stored Image Content
Systems and methods for previewing newly captured image content and reviewing previously stored image content using an electronic device are provided. The previously stored image content may include an individual distinct image or at least two consecutive video frame images.
Template-Based Excerpting and Rendering of Multimedia Performance
Disclosed herein are computer-implemented method, system, and computer-readable storage-medium embodiments for implementing template-based excerpting and rendering of multimedia performances technologies. An embodiment includes at least one computer processor configured to retrieve a first content instance and corresponding first metadata. The first content instance may include a first plurality of structural elements, with at least one structural element corresponding to at least part of the first metadata. The first content instance may be transformed by a rendering engine running on the at least one computer processor and/or transmitted to a content-playback device.
Event-triggered video creation with data augmentation
A method for creating a video that is generated based on the occurrence of pertinent events within a period of time. This video may be a summary video that includes video segments from multiple sources. The video may be augmented to display data describing pertinent events that occur.
TEMPLATE-BASED EXCERPTING AND RENDERING OF MULTIMEDIA PERFORMANCE
Disclosed herein are computer-implemented method, system, and computer-readable storage-medium embodiments for implementing template-based excerpting and rendering of multimedia performances technologies. An embodiment includes at least one computer processor configured to retrieve a first content instance and applying a template that results in transforming the first content instance. The first content instance may include a plurality of structural elements. The first content instance may be transformed by a rendering engine running on the at least one computer processor and/or transmitted to a content-playback device. An embodiment of transforming the first content instance includes trimming the content instance based on requirements provided by social media platforms.
INFORMATION PROCESSING APPARATUS, INFORMATION PROCESSING METHOD, AND PROGRAM
There is provided an information processing apparatus, an information processing method, and a program that make it possible to assist deaf and hard-of-hearing people in viewing a video when the video is being played back. The information processing apparatus includes a controller. The controller generates at least one of an oscillation signal corresponding to sound-effect caption data or an oscillation signal corresponding to vocalization caption data on the basis of a waveform of sound data using a result of analyzing caption information and sound information that are included in a video file, the sound-effect caption data being used to represent a sound effect in the form of text information, the vocalization caption data being used to represent a vocalization of a person in the form of text information, the sound-effect caption data and the vocalization caption data being included in caption data that is included in the caption information, the sound data being included in the sound information.
Audio and video recording and streaming in a three-computer booth
A system and method are presented for recording audio and video of an individual within a kiosk on separate audio and video computers that are locally connected to the kiosk. Instructions are provided to the individual through a locally connected controller computer. A remote user computer requests recorded data from the kiosk. The controller computer prompts the audio and video computers to separately stream audio and video to the remote user computer. The controller computer divides the audio and video data into time segments, and then presents different versions of the session to different users, with each different version comprising a different set of time segments. A central system server provides searching capabilities to the user computer to search and request data from a plurality of remotely located kiosks, each having separate controller, audio, and video computers.
EVENT-TRIGGERED VIDEO CREATION WITH DATA AUGMENTATION
A method for creating a video that is generated based on the occurrence of pertinent events within a period of time. This video may be a summary video that includes video segments from multiple sources. The video may be augmented to display data describing pertinent events that occur.