Patent classifications
H04N21/23418
OBJECT OR REGION OF INTEREST VIDEO PROCESSING SYSTEM AND METHOD
Systems, methods and apparatus for processing video can include a processor. The processor can be configured to perform object detection to detect visual indications of potential objects of interest in a video scene, to receive a selection of an object of interest from the potential objects of interest, and to provide enhanced video content within the video scene for the object of interest indicated by the selection.
Media manipulation using cognitive state metric analysis
Data on a user interacting with a media presentation is collected at a client device. The data includes facial image data of the user. The facial image data is analyzed to extract cognitive state content of the user. One or more emotional intensity metrics are generated. The metrics are based on the cognitive state content. The media presentation is manipulated, based on the emotional intensity metrics and the cognitive state content. An engagement score for the media presentation is provided. The engagement score is based on the emotional intensity metric. A facial expression metric and a cognitive state metric are generated for the user. The manipulating includes optimization of the previously viewed media presentation. The optimization changes various aspects of the media presentation, including the length of different portions of the media presentation, the overall length of the media presentation, character selection, music selection, advertisement placement, and brand reveal time.
METHOD FOR JUST-IN-TIME TRANSCODING OF BYTERANGE-ADDRESSABLE PARTS
A method including: ingesting a video segment and a set of video features of the video segment; estimating a part size distribution for the video segment based on the set of video features and a first rendition of the video segment; calculating a maximum expected part size based on a threshold percentile in the part size distribution; at a first time, transmitting, to an video player, a manifest file indicating a set of byterange-addressable parts of the video segment in the first rendition, each byterange addressed part characterized by the maximum expected part size; at a second time, receiving, a playback request for a first byterange-addressable part; transcoding the first byterange-addressable part; in response to the maximum expected part size exceeding a size of the first byterange-addressable part in the first rendition, appending padding data to the first byterange-addressable part; and transmitting the first byterange-addressable part to the AV player.
SYSTEMS AND METHODS FOR GENERATING METADATA FOR A LIVE MEDIA STREAM
Systems and methods are described to dynamically generate metadata for a live media stream. The system determines that a first user on a social media network has started a live media stream. In response, the system identifies a topic of the live media stream based on a frame of the live media stream and identifies another person featured in the frame of the live media stream based on social connections of the first user in the social media network. The system then generates a title for the live media stream based on the identified topic and the identified person, and transmits a notification to a second user that the first user is streaming live, where the notification includes the generated title.
AUTOMATED CONTENT IDENTIFICATION FOR BINGE WATCHING OF DIGITAL MEDIA
“Binge watching” of multiple episodes of a program is improved by the player device automatically skipping repeated portions of the program. Opening and closing credit scenes, for example, can be automatically skipped to thereby allow the viewer to progress through the entire season of programming at an even faster rate than was previously thought possible. Programming to be skipped may be identified by detecting audio or other digital fingerprints in the content itself, for example. Content to be skipped may be identified to the playback device according to presentation time stamp (PTS) or other time markers.
DETECTION DEVICE
A detection device detecting a scene related to a sponsor credit included in a commercial message from a target video is provided. The detection device comprises a detection unit that associates, from a preliminary video, a still image related to the sponsor credit with an audio signal related to the sponsor credit included other than in a frame or an audio signal configuring the commercial message so as to detect the scene related to the sponsor credit from the target video.
PRODUCING AND ADAPTING VIDEO IMAGES FOR PRESENTATION DISPLAYS WITH DIFFERENT ASPECT RATIOS
The described embodiments include systems and methods for producing and adapting images, such as video images, for presentation on display devices that have various different aspects ratios, such as 4:3, 16:9, 9:16, etc. In one embodiment, a method for producing content, such as video images, can begin by selecting an original aspect ratio and determining, within at least a first scene in the content, a position of a subject in the first scene. In one embodiment, the original aspect ratio can be substantially square (e.g., 1:1). Metadata can then be created, based on the position of the subject in the first scene, to guide playback devices to asymmetrically crop the content, relative to the position, for display on display devices that have aspect ratios that are different than the original aspect ratio. Other methods and systems are also described.
Computing system with DVE template selection and video content item generation feature
In one aspect, an example method includes (i) receiving a first group of video content items; (ii) identifying from among the first group of video content items, a second group of video content items having a threshold extent of similarity with each other; (iii) determining a quality score for each video content item of the second group; (iv) identifying from among the second group of video content items, a third group of video content items each having a quality score that exceeds a quality score threshold; and (v) based on the identifying of the third group, transmitting at least a portion of at least one video content item of the identified third group to a digital video-effect (DVE) system, wherein the system is configured for using the at least the portion of the at least one video content item of the identified third group to generate a video content item.
Systems and methods for video splicing and displaying
The present disclosure relates to a system and method for synchronous video display on at least one display. The method may comprises receiving a channel of video signal from each data acquisition port of a plurality of data acquisition ports during a time interval, each channel of video signal comprising a plurality of video frames captured during the time interval. The method may also comprises assigning a count value for each video frame of the channel of video signal as synchronization information for each video frame of the channel of video signal to form a pool of video frames each corresponding to a count value. The method may further comprises selecting video frames with the same count value from the pool of video frames as synchronized video frames, and transmitting, through the plurality of output ports, the synchronized video frames for synchronous display on the at least one display.
Video transmission method and system and device
A video transmission method, system and a device includes a sending device that generates and sends a source video stream, where the source video stream includes a plurality of video data packets, each video data packet includes discard indication information, a discard priority indicated by discard indication information included in a first video data packet is higher than a discard priority indicated by discard indication information included in a second video data packet, video data in a video frame whose frame type is a non-reference B frame is encapsulated into the first video data packet, and video data in a video frame whose frame type is an I frame, a P frame, or a reference B frame is encapsulated into the second video data packet.