H04N21/234336

SYSTEMS AND METHODS FOR REAL TIME FACT CHECKING DURING STREAMING VIEWING
20230156291 · 2023-05-18 ·

The present disclosure relates to systems and methods for fact-checking contemporary and/or live video streams. In one aspect, the disclosed systems and methods may provide fact-checked commentary and annotations to enhance the enjoyment and engagement with the streaming media. In one aspect, the systems and methods nay generate an overlay for the streaming media that includes fact-checked annotations and comments about statements or assertions made during the live streaming media.

Processing and delivery of segmented video

Self-learning systems process incoming data from sources such broadcast, cable, or IP-driven television and can discover topics that broadly describe the incoming data in real-time. These topics can be used to gather and store metadata from various metadata sources such as social networks. Using the metadata, content delivery systems working in parallel with the self-learning systems can deliver highly contextualized supplementary content to client applications, such as mobile devices used as “second screen” devices.

CROSS-MEDIA STORAGE COORDINATION

Methods and a media system and storage system for cross-media storage coordination include but are not limited to storing a first data version of specified content based on a particular media format: storing at least a second data version of related content based on a different media format: providing a cross-reference between the first data version and the at least second data version to enable coordinated management by a designated user and/or an approved device for search and possible retrieval of the first data version and/or the at least second data version: and implementing communication access by one or more parties and/or the designated user via a communication type that is correlated with the first data version and/or the at least second data version.

AUTOMATED VOICE TRANSLATION DUBBING FOR PRERECORDED VIDEO

A method for aligning a translation of original caption data with an audio portion of a video is provided. The method involves identifying original caption data for the video that includes caption character strings, identifying translated language caption data for the video that includes translated character strings associated with audio portion of the video, and mapping caption sentence fragments generated from the caption character strings to corresponding translated sentence fragments generated from the translated character strings based on timing associated with the original caption data and the translated language caption data. The method further involves estimating time intervals for individual caption sentence fragments using timing information corresponding to individual caption character strings, assigning time intervals to individual translated sentence fragments based on estimated time intervals of the individual caption sentence fragments, generating a set of translated sentences using consecutive translated sentence fragments, and aligning the set of translated sentences with the audio portion of the video using assigned time intervals of individual translated sentence fragments from corresponding translated sentences.

METHODS AND SYSTEMS FOR DETECTING AUDIO OUTPUT OF ASSOCIATED DEVICE
20230199248 · 2023-06-22 ·

Systems and methods for determining whether a first electronic device detects a media item that is to be output by a second electronic device is described herein. In some embodiments, an individual may request, using a first electronic device, that a media item be played on a second electronic device. The backend system may send first audio data representing a first response to the first electronic device, along with instructions to delay outputting the first response, as well as to continue sending audio data of additional audio captured thereby. The backend system may also send second audio data representing a second response to the second electronic device along with the media item. Text data may be generated representing the captured audio, which may then be compared with text data representing the second response to determine whether or not they match.

Methods and apparatus to determine the speed-up of media programs using speech recognition
11683558 · 2023-06-20 · ·

Methods, apparatus, systems and articles of manufacture are disclosed to determine the speed-up of media programs using speech recognition. An example apparatus disclosed herein is to perform speech recognition on a first audio clip collected by a media meter to recognize a first text string associated with the first audio clip, compare the first text string to a plurality of reference text strings associated with a corresponding plurality of reference audio clips to identify a matched one of the reference text strings, and estimate a presentation rate of the first audio clip based on a first time associated with the first audio clip and a second time associated with a first one of the reference audio clips corresponding to the matched one of the reference text strings.

Subtitle information display method and apparatus, and electronic device, and computer readable medium

A subtitle information display method includes: when an editing operation of a user for initial subtitle information of video information is detected, determining a video display region and an edited subtitle display region in an application display page; if the subtitle display region is not a subregion in the video display region, determining a first extension length and a first extension direction for each edge length of the video display region based on region information of the video display region and region information of the subtitle display region; extend the video display region within a region range corresponding to the application display page, based on the first extension length and the first extension direction, so that the extended video display region includes the subtitle display region; and displaying edited subtitle information in the subtitle display region.

Text-driven editor for audio and video assembly

The disclosed technology is a system and computer-implemented method for assembling and editing a video program from spoken words or soundbites. The disclosed technology imports source audio/video clips and any of multiple formats. Spoken audio is transcribed into searchable text. The text transcript is synchronized to the video track by timecode markers. Each spoken word corresponds to a timecode marker, which in turn corresponds to a video frame or frames. Using word processing operations and text editing functions, a user selects video segments by selecting corresponding transcribed text segments. By selecting text and arranging that text, a corresponding video program is assembled. The selected video segments are assembled on a timeline display in any chosen order by the user. The sequence of video segments may be reordered and edited, as desired, to produce a finished video program for export.

SYSTEMS AND METHODS FOR PROVIDING WORD DEFINITIONS BASED ON USER EXPOSURE
20170316708 · 2017-11-02 ·

Systems and methods are disclosed herein for providing a definition of a word encountered by the user during media consumption when a profile of the user reflects an insufficient understanding of the word. To this end, a media guidance application may detect a word appearing in media, and may determine that the word satisfies a user-specific criterion. Further, the media guidance application may determine that the exposure of the user to the word reflects an insufficient understanding of the word by calculating the level of exposure of the user to the word and comparing the level of exposure to an exposure criterion. The media guidance application may then provide a definition of the word to the user.

Content receiver system and method for providing supplemental content in translated and/or audio form
09826270 · 2017-11-21 · ·

Systems and methods utilize a content receiver to provide supplemental content, such as news content, personal content and advertising content, to a user. Received data is formatted as supplemental content by the content receiver based on user preference information stored therein, and the formatted supplemental content is transmitted to a content display device. The supplemental content is provided to the user in addition or as an alternative to video content, and may replace or supplement closed captioning content. The supplemental content may be translated into another language and/or converted into audio signals utilizing the content receiver. Systems and methods also utilize a content receiver to translate data such as text data into another language. Text data may, in addition or alternatively, be converted into audio signals utilizing the content receiver.