Patent classifications
G06F16/745
STREAMING MEDIA DATA INCLUDING AN ADDRESSABLE RESOURCE INDEX TRACK
An example device for retrieving media data includes a memory configured to store media data; and one or more processors implemented in circuitry and configured to: retrieve data of an addressable resource information (ARI) track of a media presentation, the data of the ARI track describing addressable resources and subsets of a switching set of the media presentation, the switching set comprising a plurality of media tracks including the addressable resources, the ARI track being a single index track of the media presentation, the addressable resources comprising retrievable media data; determine durations and sizes of the addressable resources from the data of the ARI track; determine one or more of the addressable resources to retrieve using the data of the ARI track including the durations and sizes of the addressable resources; retrieve the determined addressable resources; and store the retrieved addressable resources in the memory.
Multi-modal interface in a voice-activated network
Systems and methods of the present technical solution enable a multi-modal interface for voice-based devices, such as digital assistants. The solution can enable a user to interact with video and other content through a touch interface and through voice commands. In addition to inputs such as stop and play, the present solution can also automatically generate annotations for displayed video files. From the annotations, the solution can identify one or more break points that are associated with different scenes, video portions, or how-to steps in the video. The digital assistant can receive input audio signal and parse the input audio signal to identify semantic entities within the input audio signal. The digital assistant can map the identified semantic entities to the annotations to select a portion of the video that corresponds to the users request in the input audio signal.
Collection of timepoints and mapping preloaded graphs
The described technology is generally directed towards causing the generation of a content selection graph (or set of graphs) upon receipt of a notification that a new graph or replacement graph is needed with respect to a starting timepoint. For a new graph, a timepoint for that new graph indicates the need for the new graph at a given starting time. For a replacement graph, a notification (subscribed from a monitoring service) can indicate that an existing graph has changed; a replacement graph is generated with a graph identifier of the replacement graph, which then replaces the graph identifier of the existing graph in the mappings of valid graphs to start times. A Redis cache can be used to maintain the various graph sets, including the active graph sets and graph sets that will become active at a future time.
Time marking chapters in media items at a platform using machine-learning
Methods and systems for time marking of media items at a platform using machine-learning are provided herein. A media item is provided to users of a platform. An indication of the identified media item is provided as input to a machine-learning model that is trained using different feature types of historical media items to predict a plurality of content segments of a given media item each depicting, to the one or more users, a distinct section of the media item. One or more outputs of the machine-learning model are obtained comprising time marks identifying each of the plurality of content segments of the media item. Each of the plurality of content segments are associated with a segment start indicator for a timeline of the media item. The media item and an indication of each segment start indicator is provided for presentation to at least one user.
Device, method, and graphical user interface for navigating media content
An electronic device provides, to a display, while in a screensaver mode, data to present a first media, that includes a first visual motion effect. In response to receiving a user input on a remote user input device, a type of the user input on the remote user input device is determined. If the user input is of a first type, the device provides, to the display, data to present the first media, that includes the first visual motion effect, with corresponding descriptive text. If the user input is of a second type, the device exits the screensaver mode.
System and method of dynamic random access rendering
A system is provided for dynamic random access rendering of media content. The system includes a rendering tool that loads a recipe having a reference to input essence and an instruction that collectively generates an output essence using the at least one input essence. The system further includes a render engine that execute the instruction and includes a file format parser configured to load the input essence from a file of media content. The render engine also includes plugin having a web server embedded therein that is communicatively coupled with a TCP port for receiving a request from a client device for the output essence. The render engine generates the output essence from the input essence in accordance with the instruction in the recipe and transmits the generated output essence to the client device for display thereon.
System and method for generating dynamic media
The present invention provides a method for generating customized dynamic video, said method comprising the steps of: partitioning customized dynamic video's timeline, using a predefined partition scheme, into a series of chunk descriptors, each defined by an interval of the timeline and corresponding to a video chunk of the customized dynamic video; for each chunk descriptor, constructing an identifier based on a set of parameters that represent visual characteristics of the video chunk, and associating said identifier to the chunk descriptor; for each chunk descriptor, querying a cache of previously rendered and encoded video chunks for a video chunk indexed by the identifier matching that of the chunk descriptor; in the case that a video chunk whose identifier matches the queried identifier is found, retrieving said video chunk from the cache; otherwise, rendering and encoding the video chunk, for the chunk descriptor's timeline interval, and saving said video chunk in the cache, indexed by the identifier.
QUERY SYSTEM WITH SPOILER PREVENTION
Systems and methods for generating a reply to a query are provided. A query about an event in a content recording during playback of the content recording is received. A type of the event based on the query is determined. A playback position in a timeline of the content recording is determined. Based on the type of the event, an event distribution table is obtained, the table comprising one or more event identifiers and one or more corresponding occurrence times for the one or more event identifiers in the timeline of the content recording. The playback position of the content recording to the one or more occurrence times is compared. A reply to the query is generated, for aural or visual presentation, the reply being based on a result of the comparing, the reply comprising data about at least one event corresponding to the one or more event identifiers.
Determining one or more events in content
According to some aspects, methods and systems may include receiving, by a computing device, metadata identifying an event occurring in a video program, and determining an expected motion of objects in the identified event. The methods and systems may further include analyzing motion energy in the video program to identify video frames in which the event occurs, and storing information identifying the video frames in which the event occurs.
USER INTERFACE FOR LABELING, BROWSING, AND SEARCHING SEMANTIC LABELS WITHIN VIDEO
Systems and methods for browsing, searching, and/or viewing video with associated semantic labels via a user interface are presented herein. Presentation of a user interface may be effectuated. The user interface may be configured to display scenes of one or more videos and/or one or more semantic labels associated with the scenes. A first scene associated with a first video and a first semantic label associated with the first scene may be displayed in the user interface. Playback of the first video in the user interface may cause one or more changes in the display of semantic labels over time, such that responsive to playback of the first video including display of a second scene after the first scene, a second semantic label is displayed and the first semantic label is no longer displayed. The second semantic label may be associated with the second scene.