H04N21/4394

Generating videos with a character indicating a region of an image
11595738 · 2023-02-28 · ·

Methods, systems, and computer-readable media for generating videos with characters indicating regions of images are provided. For example, an image containing a first region may be received. At least one characteristic of a character may be obtained. A script containing a first segment of the script may be received. The first segment of the script may be related to the first region of the image. The at least one characteristic of a character and the script may be used to generate a video of the character presenting the script and at least part of the image, where the character visually indicates the first region of the image while presenting the first segment of the script.

Customizing search results in a multi-content source environment

Described herein are various embodiments for customizing search results in a multi-content source environment. An embodiment operates by receiving input corresponding to a search from a user and retrieving a content history indicating which content was previously viewed by the user. It is determined that the content of the content history is organized into one or more preconfigured categories. A new category of content is generated based on the content history for the user. The content of the content history for user is arranged based on both the new category and at least a subset of the one or more preconfigured categories. The arranged content is displayed in a manner customized to the user.

Systems and Methods for Assessing Viewer Engagement
20180007431 · 2018-01-04 ·

A system for quantifying viewer engagement with a video playing on a display includes at least one camera to acquire image data of a viewing area in front of the display. A microphone acquires audio data emitted by a speaker coupled to the display. The system also includes a memory to store processor-executable instructions and a processor. Upon execution of the processor-executable instructions, the processor receives the image data and the audio data and determines an identity of the video displayed on the display based on the audio data. The processor also estimates a first number of people present in the viewing area and a second number of people engaged with the video. The processor further quantifies the viewer engagement of the video based on the first number of people and the second number of people.

UNOBTRUSIVELY ENHANCING VIDEO CONTENT WITH EXTRINSIC DATA

The playback of video content upon a display is enhanced through the unobtrusive presentation of extrinsic data upon the same display. A video content feature is rendered on a display. A quantity of extrinsic data relevant to a current time in the video content feature is determined based at least in part on viewing history of a user. A graphical element presenting at least a portion of the extrinsic data is rendered on the display while the video content feature is also being rendered on the display.

PRESENTING MOBILE CONTENT BASED ON PROGRAMMING CONTEXT
20180011849 · 2018-01-11 ·

Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for generating search queries in response to obtaining audio samples on a client device. In one aspect, a method includes the actions of i) receiving audio data from a client device, ii) identifying specific content from captured media based on the received audio data, wherein the identified specific content is associated with the received audio data and the captured media includes at least one of audio media or audio-video media, iii) obtaining additional metadata associated with the identified content, iv) generating a search query based at least in part on the obtained additional metadata, and v) returning one or more search results to the client device, the one or more search results responsive to the search query and associated with the received audio data.

AUDIO TIME SYNCHRONIZATION USING PRIORITIZED SCHEDULE

A method is provided for synchronizing playback of audio an/or video by a plurality of separate devices in a computer network, e.g. in a wi-fi network. Each separate device is programmed to select a synchronization mechanism in accordance with a predetermined prioritized list of at least two different synchronization mechanisms, and to use the selected synchronization mechanism for synchronizing audio and/or video playback. E.g. use of a clock based on the audio codec clock can be set to a higher priority than use of the system clock, which provides a poorer precision. A session leader serves to provide the synchronization to other separate devices in a session, however a group of two or more separate devices within the session may agree on selecting a synchronization mechanism providing a higher precision than the one provided by the session leader. E.g. to allow high precision timing between separate left and right loudspeakers in a stereo setup. A group leader can be elected to provide synchronization to a group of devices using a higher synchronization precision than the synchronization mechanism provided by the overall session leader. E.g. a dedicated synchronization channel separate from the audio/video streaming channel may be selected.

METHODS AND APPARATUS TO IDENTIFY AND CREDIT MEDIA USING RATIOS OF MEDIA CHARACTERISTICS

Apparatus, systems, articles of manufacture, and methods to identify and credit media using ratios of media characteristics are disclosed herein. Example apparatus to identify media include at least one memory, instructions, and at least one processor to execute the instructions to: determine a first ratio based on a first time interval and a second time interval of a monitored media signal; determine a second ratio based on the second time interval and a third time interval of the monitored media signal; generate a first ratio signature based on the first and second ratios; and initiate transmission of the first ratio signature to a recipient that is to compare the first signature with a second ratio signature to identify the media.

AUDIOVISUAL COLLABORATION SYSTEM AND METHOD WITH LATENCY MANAGEMENT FOR WIDE-AREA BROADCAST AND SOCIAL MEDIA-TYPE USER INTERFACE MECHANICS

Techniques have been developed to facilitate the livestreaming of group audiovisual performances. Audiovisual performances including vocal music are captured and coordinated with performances of other users in ways that can create compelling user and listener experiences. For example, in some cases or embodiments, duets with a host performer may be supported in a sing-with-the-artist style audiovisual livestream in which aspiring vocalists request or queue particular songs for a live radio show entertainment format. The developed techniques provide a communications latency-tolerant mechanism for synchronizing vocal performances captured at geographically-separated devices (e.g., at globally-distributed, but network-connected mobile phones or tablets or at audiovisual capture devices geographically separated from a live studio).

Mini-Banner Content
20230005199 · 2023-01-05 · ·

Devices, systems, and methods are provided for use in interpreting, converting, generating, embedding, presenting, storing and otherwise using mini-banner content. For at least one embodiment, a mini-banner content system may include a secondary content system element which executes non-transient computer executable instructions to configure: a content interpreter to interpret secondary content, identify aspect information, and output the aspect information; a content converter, which when receiving aspect information at least one of retrieves a first element corresponding to the aspect information and/or generates a second element corresponding to the aspect information, and generates a mini-banner content element based upon results of the retrieving operation and/or the generating operation.

Device and method for visually displaying speaker's voice in 360-degree video

An electronic device is disclosed. In addition, various embodiments identified through the specification are possible. The electronic device includes a display, a processor, and a memory storing instructions that, when executed by the processor, cause the processor to display, when a video supporting a plurality of orientation regions is played, a screen of a first orientation region among the plurality of orientation regions and a first text corresponding to a voice of a first speaker in the screen, and display, in response to a user input of selecting a voice of a second speaker located in a second orientation region, a screen of the second orientation region.