Patent classifications
H04N21/43074
INFORMATION PROCESSING SYSTEM AND INFORMATION PROCESSING METHOD
An information processing system for obtaining an audio content file for video data providing video content representing a sport event, including: a receiver configured to receive a data stream including the video data; a preference data obtainer configured to obtain preference data, wherein the preference data indicate a selected competitor participating in the sport event; a category identifier obtainer configured to obtain a category identifier from a machine learning algorithm into which the video data is input, wherein the machine learning algorithm is trained to classify a scene represented in the video content into a category of a predetermined set of categories associated with the sport event, wherein the category identifier indicates the category into which the scene is classified; an audio content file obtainer configured to obtain, based on the obtained category identifier and the obtained preference data, the audio content file from a prestored set of audio content files, wherein the audio content file provides audio content associated with the category of the scene and the preference data; and a synchronizer configured to synchronize the audio content and the video content for synchronized play back of the scene by a media player configured to play back the video content and the audio content file.
Automated voice translation dubbing for prerecorded video
A method for aligning a translation of original caption data with an audio portion of a video is provided. The method includes identifying, by a processing device, original caption data for a video that includes a plurality of caption character strings. The processing device identifies speech recognition data that includes a plurality of generated character strings and associated timing information for each generated character string. The processing device maps the plurality of caption character strings to the plurality of generated character strings using assigned values indicative of semantic similarities between character strings. The processing device assigns timing information to the individual caption character strings based on timing information of mapped individual generated character strings. The processing device aligns a translation of the original caption data with the audio portion of the video using assigned timing information of the individual caption character strings.
Synchronizing independent media and data streams using media stream synchronization points
A messaging channel is embedded directly into a media stream. Messages delivered via the embedded messaging channel are extracted at a client media player. According to a variant embodiment, and in lieu of embedding all of the message data in the media stream, only a coordination index is injected, and the message data is sent separately and merged into the media stream downstream (at the client media player) based on the coordination index. In one example embodiment, multiple data streams (each potentially with different content intended for a particular “type” or class of user) are transmitted alongside the video stream in which the coordination index (e.g., a sequence number) has been injected into a video frame. Based on a user's service level, a particular one of the multiple data streams is released when the sequence number appears in the video frame, and the data in that stream is associated with the media.
Client-side offload of graphics effects processing
A server offloads graphics effects processing to a client device with graphics processing resources by determining a modification to a graphics effects operation, generating a portion of a rendered video stream using the modification to the graphics effects operation, and providing an encoded representation of the portion of the rendered video stream to the client device, along with metadata representing the modification implemented. The client device decodes the encoded representation to recover the portion of the rendered video stream and selectively performs a graphics effects operation on the recovered portion to at least partially revert the resulting graphics effects for the portion to the intended effects without the modification implemented by the server.
Content-modification system with determination of input-buffer switching delay feature
In one aspect, a method includes: (i) determining a baseline input-to-output delay, which represents a time period between when content is input into a first input buffer and output by an output buffer; (ii) establishing a synchronous lock between (a) first fingerprint data representing content in the first input buffer and (b) second fingerprint data representing content in the output buffer; (iii) determining an instruction time-point at which a content-presentation device is instructed to switch from using content in the first input buffer to populate the output buffer, to using content in the second input buffer to populate the output buffer; (iv) determining a loss of synchronous lock time-point; (v) using the determined baseline input-to-output delay, the determined instruction time-point, and the determined loss of synchronous lock time-point to determine an input-buffer switching delay; and (vi) using the determined input-buffer switching delay to facilitate performing a content-modification operation.
System and method for synchronizing on-air commercial programming with interactive applications
There is provided a system and method for the synchronizing on-air commercial programming with interactive applications, providing interactive content synchronized with on-air television programming. There is provided a system comprising a client device with a display device and an input device and running a synchronization application. The client device receives digital data transmitted from a server, and transmits user feedback information generated from user interactions with the digital data back to the server. The digital data are interactive advertisements, such as online social games, that increase users' awareness of the advertised products or messages. The digital data are delivered to the user's client device in synchronization with the broadcasting of on-air television commercials.
Method for synchronizing an additional signal to a primary signal
The present invention relates to a method for synchronizing an additional signal to a primary signal. Synchronization information for a primary signal is generated by extracting at least one signal feature sequence of the primary signal and comparing it to DB feature sequences stored in a database. If the signal feature sequence matches one of the DB feature sequences to a predetermined degree, then synchronization information of the matching DB feature sequences is allocated to the primary signal at a position specified by the signal feature sequence. The synchronization information is transmitted to a playback device, which outputs an additional signal to the primary signal based on the synchronization information.
Multi-viewpoint multi-user audio user experience
An apparatus including circuitry configured for receiving a spatial media content file including a plurality of viewpoints; circuitry configured for determining a first viewpoint from the plurality of viewpoints for a first user consuming the spatial media content file; circuitry configured for receiving an indication that affects an audio rendering of the first viewpoint for the first user, wherein the indication is associated with one or more actions of at least one second user consuming the spatial media content file; and circuitry configured for controlling the audio rendering of the first viewpoint for the first user in response to the receiving of the indication based on at least one of: a position and/or orientation of the first user, and the one or more actions of the second user.
GENERATING VERIFIED CONTENT PROFILES FOR USER GENERATED CONTENT
Systems and methods for searching, identifying, scoring, and providing access to companion media assets for a primary media asset are disclosed. In response to a request for companion content, metadata within a predefined time period of a play position when the request was made, is downloaded. A dynamic search template that contains search parameters based on the downloaded metadata is generated. In response to the search conducted using the search template, a plurality of companion media assets are identified and then verified. A trust score for the companion media asset is accessed. The trust score may be analyzed and modified based on its contextual relationship to the play position of the primary media asset. If the trust score is within a rating range, then a link to access the companion media asset, or a specific segment or play position within the companion media asset, is provided.
Synchronizing video content with extrinsic data
Disclosed are various embodiments for associating and synchronizing extrinsic data with video content at particular points of time in the video content. In one embodiment, an application identifies a video content feature currently being presented via a display device. A current time in the video content feature are determined. Performers potentially depicted in the video content feature at the current time are determined. The application then generates a user interface configured to receive a user selection of a particular performer from the performers, where the user selection indicates that the particular performer is depicted in the video content feature at the current time.