H04N21/234336

CAPTION SERVICE SYSTEM FOR REMOTE SPEECH RECOGNITION

The present invention provides a caption service system for remote speech recognition, which provides caption service for the hearing impaired. This system includes a speaker and a live broadcast equipment at A, a listener-typist and a computer at B, a hearing impaired and a live screen at C, and an automatic speech recognition (ASR) caption server at D. Connect the live broadcast equipment, the computer, the live screen and the ASR caption server with a network. The speaker's audio is sent to the automatic speech recognition (ASR) caption server to be converted into text, which is corrected by the listener-typist, and then the text caption is sent to the live screen of the hearing impaired together with the speaker's video and audio, so that the hearing impaired can see the text caption spoken by the speaker.

LIVESTREAM VIDEO IDENTIFICATION

A computing system is described herein, where the computing system is configured to perform a search over a computer-readable index based upon a query for a user. The computer-readable index includes an identifier for a livestream video that is currently being livestreamed by way of a livestreaming service and values for respective attributes of the livestream video. The values for the respective attributes are updated as content of the livestream video alters over time. The livestream video is identified from amongst several livestream videos based upon the search, where the video is identified due to a set of values specified in the query corresponding to the values for the respective attributes in the computer-readable index. Upon the livestream video being identified, an identifier of the livestream video is transmitted to a client computing device of the user.

Real Time Popularity Based Audible Content Acquisition

A personalized news service provides personalized news programs for its users by generating personalized combinations of audible versions of news stories derived from text-based versions of the news stories. The audible versions may be generated from the text-based version by a text-to-speech system, or may by recording a person reading aloud the text-based version. To acquire recordings, the personalized news service can make a determination that a particular news story has a threshold extent of popularity. The news service can then transmit a request to a remote recording station for a recording of a verbal reading of the particular news story. The news service can then receive the requested recording from the remote recording station.

METHOD AND SYSTEM FOR REAL-TIME TRANSCODING OF MPEG-DASH ON-DEMAND MEDIA SEGMENTS WHILE IN TRANSIT FROM CONTENT HOST TO DASH CLIENT

A system, method and computer program product for real-time post-processing system that transforms MPEG-DASH on-demand media streams, including a DASH media player device; an intercepting media server device; a MPEG-DASH content origin server device; and a proxy media client device coupled to the DASH media player device and the intercepting media server device and configured to intercept MPEG-DASH HTTP requests from the DASH media player device and forward the intercepted requests to the intercepting media server device instead of the MPEG-DASH content origin server device. The intercepting media server device is configured to act as an HTTP proxy device, and forward the intercepted requests to the MPEG-DASH content origin server, and with each corresponding MPEG-DASH media subsegment acquired perform analysis of the video media content within the subsegment and apply selective transcoding.

Content providing server, content providing terminal, and content providing method

A method for operating a user terminal providing a video slide service includes uploading a UGC video to a server; receiving, from the server, scene metadata information for each reproduction section corresponding to the UGC video; generating a video slide file based on the scene metadata information for each reproduction section; displaying an item corresponding to the video slide file; and displaying a page screen composed of representative image information and subtitle information, for each reproduction section of the UGC video in response to a user selection on the item.

CONTENT ACCESS DEVICES THAT USE LOCAL AUDIO TRANSLATION FOR CONTENT PRESENTATION
20230095557 · 2023-03-30 ·

A content access device uses local audio translation for content presentation. The content access device receives video and first audio data associated with a first language. The content access device uses translation software and/or other automated translation services to translate the first audio data to second audio data associated with a second language. The content access device synchronizes the video with the second audio data and outputs the video and the second audio data for presentation. The first audio data may be audio, text, and so on. The second audio data may be output as audio, text, and so on.

VIRTUAL VIDEO LIVE STREAMING PROCESSING METHOD AND APPARATUS, STORAGE MEDIUM AND ELECTRONIC DEVICE

The application provides a virtual video live streaming processing method and apparatus, an electronic device, and a computer-readable storage medium, and relates to the field of virtual video live streaming technologies. The virtual video live streaming processing method includes: obtaining text data and determining to-be-synthesized video data corresponding to the text data; synthesizing a live video stream in real time according to the to-be-synthesized video data and pushing the live video stream to a live streaming client; determining target video data from the to-be-synthesized video data that has not been synthesized into a live video stream in response to a live streaming interruption request during receiving a live streaming interruption request; and synthesizing an interruption transition video stream according to the target video data and pushing the interruption transition video stream to the live streaming client. When a live video is interrupted during a virtual video live streaming process, this application may implement a smooth transition process between a current video action and a next video action without affecting real-time performance of the live video.

Video processing for embedded information card localization and content extraction
11615621 · 2023-03-28 · ·

Metadata for one or more highlights of a video stream may be extracted from one or more card images embedded in the video stream. The highlights may be segments of the video stream, such as a broadcast of a sporting event, that are of particular interest. According to one method, video frames of the video stream are stored. One or more information cards embedded in a decoded video frame may be detected by analyzing one or more predetermined video frame regions. Image segmentation, edge detection, and/or closed contour identification may then be performed on identified video frame region(s). Further processing may include obtaining a minimum rectangular perimeter area enclosing all remaining segments, which may then be further processed to determine precise boundaries of information card(s). The card image(s) may be analyzed to obtain metadata, which may be stored in association with at least one of the video frames.

Method, system, and non-transitory computer-readable record medium for providing animation using sprite JPEG
11615573 · 2023-03-28 · ·

Disclosed is a method, system, and non-transitory computer-readable record medium for providing an animation function using a sprite joint photographic experts group (JPEG) image. An animated image providing method includes creating, by processing circuitry, a sprite joint photographic experts group (JPEG) image from a video in an animated format, and providing, by the processing circuitry, an animated thumbnail of the video based on the sprite JPEG image.

SUBTITLE INFORMATION DISPLAY METHOD AND APPARATUS, AND ELECTRONIC DEVICE, AND COMPUTER READABLE MEDIUM
20220353586 · 2022-11-03 ·

A subtitle information display method includes: when an editing operation of a user for initial subtitle information of video information is detected, determining a video display region and an edited subtitle display region in an application display page; if the subtitle display region is not a subregion in the video display region, determining a first extension length and a first extension direction for each edge length of the video display region based on region information of the video display region and region information of the subtitle display region; extend the video display region within a region range corresponding to the application display page, based on the first extension length and the first extension direction, so that the extended video display region includes the subtitle display region; and displaying edited subtitle information in the subtitle display region.