H04N9/80

Acquisition of extracts of a multimedia stream on a terminal
11470407 · 2022-10-11 · ·

A method for acquiring an extract of a multimedia stream being broadcast on a user's display device, the method being implemented in an user terminal separate from the display device and comprising the following steps: upon reception of a user input for receiving an extract of the multimedia stream being broadcast on the display device, sending to the display device a first request for an identifier of the multimedia stream being broadcast; upon reception of the identifier of the multimedia stream being broadcast, sending to the content server a second request for receiving the multimedia stream extract, said second request including a timestamp of the user input, the identifier of the multimedia stream being broadcast and a duration of the multimedia stream extract; and receiving from the content server the multimedia stream extract or a link to obtain the latter.

Stereoscopic viewer

Features for a lightweight stereoscopic viewing client are described. The client can generate accurate ground point coordinates from selections within the lightweight viewer by accumulating the transformations from original image sources to the images used to render the stereoscopic scene to accurately predict error for a point selection. The viewer may also be decoupled from a permanent image store allowing on-demand retrieval of images via a network for stereoscopic viewing.

Method and apparatus for extracting video clip

The present disclosure discloses a method and apparatus for extracting a video clip, relates to the field of artificial intelligence technology such as video processing, audio processing, and cloud computing. The method includes: acquiring a video, and extracting an audio stream in the video; determining a confidence that audio data in each preset period in the audio stream comprises a preset feature; and extracting a target video clip corresponding to a location of a target audio clip in the video; wherein the target audio clip is an audio clip within a continuous preset period, and has a confidence that the audio data includes the preset feature, which is larger than a preset confidence threshold. This method may improve the accuracy of extracting a video clip.

Information processing apparatus, program, and information processing method
09854220 · 2017-12-26 · ·

An information processing apparatus includes a content reproduction unit, a content part specification unit, a clustering unit, a class detection unit, a common class extraction unit, and a content retrieval unit. The content reproduction unit is configured to reproduce a reproduction content. The content part specification unit is configured to specify a plurality of content parts included in the reproduction content. The clustering unit is configured to classify the plurality of content parts into a plurality of clusters. The class detection unit is configured to detect a class from the plurality of content parts included in each of the plurality of clusters. The common class extraction unit is configured to extract a common class common to the plurality of clusters from the classes detected by the class detection unit. The content retrieval unit is configured to retrieve a content corresponding to the common class.

Elastic cloud video editing and multimedia search

Technologies for cloud-based media search and editing include a video editor configured to build a media query and associate the media query with a dynamic content slot of a media program. When generating video output based on the media program, the video editor transmits the media query to a cloud analytics server and receives 5 search results identifying one or more media clips produced by a number of mobile computing devices. The video editor may display a list of clips for selection by the user or may automatically include one of the clips in the output. The cloud analytics server transmits an acceptance policy defining criteria for acceptable media, based on the media query, to the mobile computing devices. The mobile computing devices configure 10 capture settings according to the acceptance policy and may display a user interface to assist the user in capturing acceptable media. Other embodiments are described and claimed.

Localized audio source extraction from video recordings

Technologies are generally described for a system to process a collection of video recordings of a scene to extract and localize audio sources for the audio data. According to some examples, video recordings captured by mobile devices from different perspectives may be uploaded to a central database. Video segments capturing an overlapping portion of the scene at an overlapping time may be identified, and a relative location of each of the video capturing devices may be determined. Audio data for the video segments may be indexed with a sub-frame time reference and relative locations as a function of overlapping time. Using the indices that include the sub-frame time references and relative locations, audio sources for the audio data may be extracted and localized. The extracted audio sources may be transcribed and indexed to enable searching, and may be added back to each video recording as a separate audio channel.

Randomly accessible visual information recording medium and recording method, and reproducing device and reproducing method

Access point pictures designated as randomly accessible positions are I pictures or P pictures. Information indicating the decoding sequence (I1, P1, B1, B2, B3, B4, P2, . . . ) of pictures functioning as access points and attribute information (picture_type) indicating whether a picture functions as an access point or is necessary for decoding of the access point following a given access point are recorded on the video information recording medium. Random access is possible even if the GOP interval is lengthened.

Digital camera user interface for video trimming

A digital video camera comprising: user controls enabling a user to select between at least an up input, a down input, a left input, a right input, and a confirmation input; and a program memory storing instructions to implement a method for trimming a digital video sequence. The method includes: selecting a digital video sequence; initiating a trimming operation; accepting user input to select a start frame and an end frame for a trimmed digital video sequence, wherein the up input and the down input are used to select between a start frame selection mode and an end frame selection mode, and the left input and the right input are used to scroll through the frames of the selected digital video sequence; and trimming the selected video sequence to include the frames between the selected start frame and the selected end frame.

Media playback method and device
09848238 · 2017-12-19 · ·

A media playback method and device that relate to the field of computer technologies are provided. The method includes turning on a camera to detect the number of human eyes within a shooting range to determine an initial number of viewers when a user performs a media playback activity, detecting, at a set interval, the number of human eyes to determine a current number of viewers, and encrypting a playback image when the current number of viewers is greater than the initial number of viewers. The method is applicable when a user views information, such as a text, a picture, or a video, on a device such as a mobile phone terminal, a pad device, or a computer.

Method and device for bounding an object in a video

The invention relates to a method for bounding an object in a video sequence F.sub.x,y,t. The method includes obtaining a subset of pixels located in the object to annotate, in each frame of the video sequence. Spatio-temporal slicing is performed on the video sequence F.sub.x,y,t, centered on the obtained subsets of pixels, resulting in a first image F.sub.y,t obtained by an horizontal concatenation of first slices, comprising the obtained subsets of pixels, and resulting in a second image F.sub.x,t obtained by a vertical concatenation of second slices. A trajectory of the obtained subsets of pixels is displayed on both the first F.sub.y,t and second F.sub.x,t image. A bounding form around the object to annotate is obtained out of four points in each frame of the video sequence, wherein the coordinates of the four points of a frame t are obtained from the coordinates of the points located in the first and second boundary of the first and second image for that frame t.