G06F16/7834

Continuous video generation from voice data

One example method includes capturing audio data at a client engine while outputting an output video, the output video being based upon an original video stored at the client engine, delivering the captured audio data to a prediction engine upon the captured audio data being captured for a pre-determined time, receiving from the prediction engine substitute frame data used by the client engine to stitch one or more frames into the original video stored at the client engine, and following stitching the one or more frames into the output video to generate an altered output video, outputting the captured audio data and the altered video from the client engine.

PRESENTING MOBILE CONTENT BASED ON PROGRAMMING CONTEXT
20180011849 · 2018-01-11 ·

Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for generating search queries in response to obtaining audio samples on a client device. In one aspect, a method includes the actions of i) receiving audio data from a client device, ii) identifying specific content from captured media based on the received audio data, wherein the identified specific content is associated with the received audio data and the captured media includes at least one of audio media or audio-video media, iii) obtaining additional metadata associated with the identified content, iv) generating a search query based at least in part on the obtained additional metadata, and v) returning one or more search results to the client device, the one or more search results responsive to the search query and associated with the received audio data.

Generation, administration and analysis of user experience testing

Systems and methods for generating, administering and analyzing a user experience study are provided. In particular, intents can be generated from a user experience study by applying one or more screener questions to participants and subjecting the screened participants to one or more tasks. Corresponding clickstreams and success data for each participant engaging in the tasks can be recorded. The success and clickstream data can also be aggregated for all the screened participants as aggregated results. Video data including audio for each of the screened participants can also be recorded.

EFFICIENT EXPLORER FOR RECORDED MEETINGS
20230029278 · 2023-01-26 ·

One example method includes generating a searchable video library. Video files are processed to extract text corresponding to the speech and to the images. The extracted text is semantically searched such that specific portions or locations of video files can be identified and returned in response to a query.

Audio transitions when streaming audiovisual media titles
11700415 · 2023-07-11 · ·

A playback application is configured to analyze audio frames associated with transitions between segments within a media title to identify one or more portions of extraneous audio. The playback application is configured to analyze the one or more portions of extraneous audio and then determine which of the one or more corresponding audio frames should be dropped. In doing so, the playback application can analyze a topology associated with the media title to determine whether any specific portions of extraneous audio are to be played outside of a logical ordering of audio samples set forth in the topology. These specific portions of extraneous audio are preferentially removed.

GENERATING VERIFIED CONTENT PROFILES FOR USER GENERATED CONTENT

Systems and methods for searching, identifying, scoring, and providing access to companion media assets for a primary media asset are disclosed. In response to a request for companion content, metadata within a predefined time period of a play position when the request was made, is downloaded. A dynamic search template that contains search parameters based on the downloaded metadata is generated. In response to the search conducted using the search template, a plurality of companion media assets are identified and then verified. A trust score for the companion media asset is accessed. The trust score may be analyzed and modified based on its contextual relationship to the play position of the primary media asset. If the trust score is within a rating range, then a link to access the companion media asset, or a specific segment or play position within the companion media asset, is provided.

Video broadcasting through selected video hosts
11553229 · 2023-01-10 ·

A method for providing captured video to a subsequent user device, via a desired video host, including at least some of allowing a user to designate, via a mobile device, at least one desired video host; allowing the subsequent user, via the subsequent user device, to be associated with the at least one desired video host; allowing the user to capture video, via the mobile device, and upload or stream the captured video to the desired video host device(s), wherein the captured video includes at least one categorization for the captured video, as designated by the user prior to capturing the video; and allowing the subsequent user to access, via the subsequent user device, the captured video, via the video host device associated with the at least one desired video host, wherein the captured video is accessed, based on the at least one categorization for the captured video.

METHOD AND DEVICE FOR VIEWING CONFERENCE
20230006851 · 2023-01-05 · ·

A method and a device for viewing a conference are provided. In the method, after a wide-view video of a specific conference, related conference event data, and speech content of each participant are obtained, a highlight video of the specific conference is correspondingly generated. Accordingly, the efficiency of conference viewing is improved.

Answering questions during video playback
11544590 · 2023-01-03 · ·

In implementations of answering questions during video playback, a video system can receive a question related to a video at a timepoint of the video during playback of the video, and determine audio sentences of the video that occur within a segment of the video that includes the timepoint. The video system can generate a classification vector from words of the question and the audio sentences, and determine an answer to the question utilizing the classification vector. The video system can obtain answer candidates, and the answer to the question can be selected as one of the answer candidates based on matching the classification vector to one of the answer vectors.

VIDEO PROCESSING OPTIMIZATION AND CONTENT SEARCHING

Techniques are disclosed for automatic scene detection and character extraction. In one example, audiovisual content with video frames, an audio recording, and timing information is received. A score, based on the frame's visual characteristics, is determined for a first frame and subsequent frames. The first frame's score and subsequent frame's scores are compared to determine if the difference between the scores is above a threshold. When the difference in scores is above a threshold, the subsequent frame is classified as a new scene. The audiovisual content is segmented into scenes and textual characters are identified in at least one frame from each scene. The characters are stored and indexed in a searchable database with the timing information for the scene where the characters were identified. The audio recording is transcribed and the transcribed words are stored and indexed in the searchable database with timing information.