Patent classifications
G06F16/7834
METHODS, SYSTEMS, AND APPARATUSES TO RESPOND TO VOICE REQUESTS TO PLAY DESIRED VIDEO CLIPS IN STREAMED MEDIA BASED ON MATCHED CLOSE CAPTION AND SUB-TITLE TEXT
Methods, Systems, and Apparatuses are described to implement voice search in media content for requesting media content of a video clip of a scene contained in the media content streamed to the client device; for capturing the voice request for the media content of the video clip to display at the client device wherein the streamed media content is a selected video streamed from a video source; for applying a NLP solution to convert the voice request to text for matching to a set of one or more words contained in at least close caption text of the selected video; for associating matched words to close caption text with a start index and an end index of the video clip contained in the selected video; and for streaming the video clip to the client device based on the start index and the end index associated with matched closed caption text.
AUDIO RECOMMENDATION BASED ON TEXT INFORMATION AND VIDEO CONTENT
An electronic device and method for audio recommendation and generation are disclosed. The electronic device receives textual information that indicates a plurality of scenes for video content, and determines a first plurality of features for the plurality of scenes. The electronic device determines a set of positions in the textual information based on the determined first plurality of features. A set of audio files are to be inserted at the set of positions related to a set of scenes of the plurality of scenes. The electronic device determines, by an artificial intelligent (AI) engine, the set of audio files for the set of scenes, based on a second plurality of features and the first plurality of features related to the set of scenes. The electronic device controls a display device to display first information corresponding to the set of positions and second information corresponding to the set of audio files.
Generating verified content profiles for user generated content
Systems and methods for searching, identifying, scoring, and providing access to companion media assets for a primary media asset are disclosed. In response to a request for companion content, metadata within a predefined time period of a play position when the request was made, is downloaded. A dynamic search template that contains search parameters based on the downloaded metadata is generated. In response to the search conducted using the search template, a plurality of companion media assets are identified and then verified. A trust score for the companion media asset is accessed. The trust score may be analyzed and modified based on its contextual relationship to the play position of the primary media asset. If the trust score is within a rating range, then a link to access the companion media asset, or a specific segment or play position within the companion media asset, is provided.
Insertion of directed content into a video asset
Technologies are provided for insertion of directed content into video assets based on analysis of audio data corresponding to the video assets. Some embodiments include a computing system that can receive audio data corresponding to a video asset. The computing system can generate, using the audio data, a time series of tags corresponding to speech conveyed by the video asset. The computing system can then generate a time-asset mapping between time and directed content using the time series and a correlation policy. The directed content can include digital media intended for a defined audience, for example. The time-asset mapping associates groups of directed content assets to respective specific times in the time series. The computing system can insert, using the mapping, a defined directed content asset from a group of directed content assets identified in the mapping.
NFT production from feature films including spoken lines
Methods and processes for manufacture of an image product from a digital image. An object in the digital image is detected and recognized. Object metadata is assigned to the object, the object metadata linking sound to the object in the digital image which produced the sound. At least one cryptographic hash of the object metadata is generated, and the hash is written to a node of a transaction processing network.
Video-based conversational interface
In an answer view, a first video segment is selected based on a first natural language input and displayed in a main display region, and a self-video image of a user is displayed in a peripheral display region having a smaller area than the main display region. To transition from the answer view to a question view, the self-video image is enlarged to replace the first video segment in the main display region. A second natural language input is received. To transition from the question view to the answer view, the self-video image is reduced to occupy the peripheral display region and the self-video image is replaced in the main display region with a second video segment selected based on the second natural language input. The video segments are pre-recorded video response segments spoken by the same person. Enlarging the self-video image masks the transition between the video segments.
PLAYBACK OF AUDIO CONTENT ALONG WITH ASSOCIATED NON-STATIC MEDIA CONTENT
This disclosure concerns the provision of media, and more particularly streaming of media. In particular, one aspect herein relates to a method performed by a server system of streaming an audio content item to an electronic device. In response to receiving a request message from the electronic device, a selected audio content item is retrieved from a first storage. Descriptive metadata including an origin-ID associated with the retrieved audio content item is determined. A second storage is browsed utilizing said metadata including the origin-ID to locate non-static media content item(s) associated with the origin-ID. In response to finding a non-static media content item associated with the origin ID, the selected audio content item is sent along with the located non-static media content item to the electronic device for simultaneous presentation of the audio content item and the located non-static media content item.
VOICE PACKET RECOMMENDATION METHOD AND APPARATUS, DEVICE AND STORAGE MEDIUM
Provided are a voice packet recommendation method and apparatus, a device and a storage medium. The method includes selecting at least one target display video for a user from among candidate display videos associated with voice packets and using voice packets to which the at least one target display video belongs as candidate voice packets; selecting a target voice packet for the user from among the candidate voice packets according to attribute information of the candidate voice packets and attribute information of the at least one target display video; and recommending the target voice packet to the user.
GENERATING VISUAL DATA STORIES
This disclosure describes one or more embodiments of systems, non-transitory computer-readable media, and methods that intelligently and automatically analyze input data and generate visual data stories depicting graphical visualizations from data insights determined from the input data. For example, the disclosed systems automatically extract data insights utilizing an in-depth statistical analysis of dataset groups from data-attribute categories within the input data. Based on the data insights, the disclosed systems can automatically generate exportable visual data stories to visualize the data insights, provide textual or audio-based natural language summaries of the data insights, and animate such data insights in videos. In some embodiments, the disclosed systems generate a visual-data-story graph comprising nodes representing visual data stories and edges representing similarities between the visual data stories. Based on the visual-data-story graph, the disclosed systems can select a relevant visual data story to display on a graphical user interface.
Video generating method, apparatus, electronic device, and computer-readable medium
A video generating method, an apparatus, an electronic device, and a computer-readable medium are provided. The method includes: acquiring a first video set and an audio material; determining a first music point of the audio material according to an amount of the video materials in the first video set; generating, according to a sorting order of the video materials in the first video set, one video clip for each first music clip in the audio material by respectively using one video material, so as to obtain a first video sequence; adjusting, in response to detecting an editing operation on the video clip in the first video sequence, the video clip in the first video sequence, so as to obtain a second video sequence; and splicing together video clips in the second video sequence, and adding the audio material as a video audio track to obtain a composite video.