H04N2017/008

Using Closed-Captioning Data To Output An Alert Indicating A Functional State Of A Back-Up Video-Broadcast System
20170230726 · 2017-08-10 ·

In one aspect, an example method for outputting an alert indicating a functional state of a back-up video-broadcast system involves a computing device receiving first closed-captioning data that corresponds to a first video-stream; the computing device receiving second closed-captioning data that corresponds to a second video-stream; the computing device making a determination that the received first closed-captioning data and the received second closed-captioning data lack a threshold extent of similarity; and responsive to the determination that the received first closed-captioning data and the received second closed-captioning data lack the threshold extent of similarity, the computing device outputting an alert.

Automatic extraction of closed caption data from frames of an audio video (AV) stream using image clipping

Exemplary methods of extracting closed caption (CC) image from a frame of an audio video (AV) stream are described. For all pixels of the frame, setting a color value of the pixels to a determined pixel value when the color value of the pixel is different from a background color value associated with CC image. A set edges is analyzed to identify one or more polygons. A polygon that contains text is determined from the one or more polygons. The frame is cropped along the polygon to obtain a CC image. Upon determination that the CC image is identical to another closed caption image a frame count associated with the other closed caption image is increased by 1; and upon determination that the CC image is not identical to the other CC image the closed caption image is stored along with a position and a time value as metadata information.

Using closed-captioning data to output an alert indicating a functional state of a back-up video-broadcast system

In one aspect, an example method for outputting an alert indicating a functional state of a back-up video-broadcast system involves a computing device receiving first closed-captioning data that corresponds to a first video-stream; the computing device receiving second closed-captioning data that corresponds to a second video-stream; the computing device making a determination that the received first closed-captioning data and the received second closed-captioning data lack a threshold extent of similarity; and responsive to the determination that the received first closed-captioning data and the received second closed-captioning data lack the threshold extent of similarity, the computing device outputting an alert.

Automatic extraction of closed caption data from frames of an audio video (AV) stream using image filtering

Exemplary methods of extracting closed caption images from frames of an audio video (AV) stream are described. A first set of frames of a first AV stream including CC images and a second set of frames not including the CC images are received. Each pixel in the first frame is replaced with a dummy pixel, upon determination that a pixel at a corresponding position in the corresponding frame has a same color value to generate a filtered frame including dummy pixels and non-dummy pixels. First coordinates of the top-left most pixel of the first frame that is not a dummy pixel and second coordinates of the bottom-right most pixel of the first frame that is not a dummy pixel are determined. The filtered frame is cropped along the first and second coordinates to extract the CC image.

Method and system for testing closed caption content of video assets
09620118 · 2017-04-11 · ·

A method and system for monitoring video assets provided by a multimedia content distribution network includes testing closed captions provided in output video signals. A video and audio portion of a video signal are acquired during a time period that a closed caption occurs. A first text string is extracted from a text portion of a video image, while a second text string is extracted from speech content in the audio portion. A degree of matching between the strings is evaluated based on a threshold to determine when a caption error occurs. Various operations may be performed when the caption error occurs, including logging caption error data and sending notifications of the caption error.

Testing rendering of screen objects
12273581 · 2025-04-08 · ·

The present disclosure relates to methods and devices for testing video data being rendered at or using a media device. A plurality of video frames to be rendered is received, each frame comprising one or more primary screen objects and at least one further screen object. The received frames are rendered at or using the media device wherein the at least one further screen object is superimposed on the one or more primary screen objects of a given frame during rendering. The rendered frames are provided to a data model. Extracted metadata indicating the presence or absence of further screen objects in the rendered video frames is the output of the data model. The data model is also provided with original metadata associated with the video frames prior to rendering. The rendering of each further screen object is then tested based on the original metadata and extracted metadata relating to a given video frame. The disclosure also extends to associated methods and devices for generating training data for testing rendering of video frame and training a data model using the training data.

TESTING RENDERING OF SCREEN OBJECTS
20250317616 · 2025-10-09 · ·

The present disclosure relates to methods and devices for testing video data being rendered at or using a media device. A plurality of video frames to be rendered is received, each frame comprising one or more primary screen objects and at least one further screen object. The received frames are rendered at or using the media device wherein the at least one further screen object is superimposed on the one or more primary screen objects of a given frame during rendering. The rendered frames are provided to a data model. Extracted metadata indicating the presence or absence of further screen objects in the rendered video frames is the output of the data model. The data model is also provided with original metadata associated with the video frames prior to rendering. The rendering of each further screen object is then tested based on the original metadata and extracted metadata relating to a given video frame. The disclosure also extends to associated methods and devices for generating training data for testing rendering of video frame and training a data model using the training data.