Patent classifications
H04N19/137
Methods and apparatus for encoding, communicating and/or using images
Methods and apparatus for capturing, communicating and using image data to support virtual reality experiences are described. Images, e.g., frames, are captured at a high resolution but lower frame rate than is used for playback. Interpolation is applied to captured frames to generate interpolated frames. Captured frames, along with interpolated frame information, are communicated to the playback device. The combination of captured and interpolated frames correspond to a second frame playback rate which is higher than the image capture rate. Cameras operate at a high image resolution but slower frame rate than images could be captured with the same cameras at a lower resolution. Interpolation is performed prior to delivery to the user device with segments to be interpolated being selected based on motion and/or lens FOV information. A relatively small amount of interpolated frame data is communicated compared to captured frame data for efficient bandwidth use.
Controlling Deblocking Filtering At Different Levels In Coded Video
Methods and apparatus for video processing are disclosed. The processing may include video encoding, video decoding, or video transcoding. One example method includes performing a conversion between a video including a picture including one or more slices and a bitstream of the video. The conversion conforms to a rule that specifies that whether a deblocking filter is applied to the one or more slices referring to a picture parameter set is based at least on a first syntax element included in the picture parameter set. The first syntax element indicates whether the deblocking filter is disabled for the picture.
Prediction refinement for combined inter intra prediction mode
A method for processing a video includes performing a conversion between a current block of visual media data and a corresponding coded representation of the visual media data, wherein the conversion of the current block includes determining whether a use of one or both of a bi-directional optical flow (BIO) technique or a decoder-side motion vector refinement (DMVR) technique to the current block is enabled or disabled, and wherein the determining the use of the BIO technique or the DMVR technique is based on a cost criterion associated with the current block.
Prediction refinement for combined inter intra prediction mode
A method for processing a video includes performing a conversion between a current block of visual media data and a corresponding coded representation of the visual media data, wherein the conversion of the current block includes determining whether a use of one or both of a bi-directional optical flow (BIO) technique or a decoder-side motion vector refinement (DMVR) technique to the current block is enabled or disabled, and wherein the determining the use of the BIO technique or the DMVR technique is based on a cost criterion associated with the current block.
Video signal processing method and device using motion compensation
Disclosed is a video signal processing method comprising the steps of: acquiring a merge index which indicates a candidate for prediction of a current block among a plurality of candidates for the prediction of the current block, which are included in a merge candidate list; acquiring motion information of the current block on the basis of motion information corresponding to the candidate indicated by the merge index; when the motion information of the current block includes a plurality of motion vectors corresponding to a list of different reference pictures, comparing picture order count (POC) differences between each of the reference pictures corresponding to the plurality of motion vectors and the current picture including the current block; correcting the plurality of motion vectors on the basis of a result of the comparison; and reconstructing the current block on the basis of the plurality of corrected motion vectors.
Video signal processing method and device using motion compensation
Disclosed is a video signal processing method comprising the steps of: acquiring a merge index which indicates a candidate for prediction of a current block among a plurality of candidates for the prediction of the current block, which are included in a merge candidate list; acquiring motion information of the current block on the basis of motion information corresponding to the candidate indicated by the merge index; when the motion information of the current block includes a plurality of motion vectors corresponding to a list of different reference pictures, comparing picture order count (POC) differences between each of the reference pictures corresponding to the plurality of motion vectors and the current picture including the current block; correcting the plurality of motion vectors on the basis of a result of the comparison; and reconstructing the current block on the basis of the plurality of corrected motion vectors.
Event/object-of-interest centric timelapse video generation on camera device with the assistance of neural network input
An apparatus including an interface and a processor. The interface may be configured to receive pixel data generated by a capture device. The processor may be configured to generate video frames in response to the pixel data, perform computer vision operations on the video frames to detect objects, perform a classification of the objects detected based on characteristics of the objects, determine whether the classification of the objects corresponds to a user-defined event and generate encoded video frames from the video frames. The encoded video frames may be communicated to a cloud storage service. The encoded video frames may comprise a first sample of the video frames selected at a first rate when the user-defined event is not detected and a second sample of the video frames selected at a second rate while the user-defined event is detected. The second rate may be greater than the first rate.
Method and apparatus for processing video signals using affine prediction
Disclosed are a method for processing video signals and an apparatus therefor. Specifically, the method for processing video signals on the basis of inter prediction comprises the steps of: inducing a motion vector predictor using motion data of a neighboring block of the current block; parsing layer data indicating the current layer to which a motion vector difference used in inter prediction of the current block belongs, in a previously defined layer structure in which the combination of one or more horizontal and vertical components of motion vector differences are divided into multiple layers; parsing index data indicating a particular combination within the current layer; inducing a motion vector difference of the current block using the layer data and index data; and inducing the motion vector for the current block by adding the motion vector difference to the motion vector predictor.
Method and apparatus for processing video signals using affine prediction
Disclosed are a method for processing video signals and an apparatus therefor. Specifically, the method for processing video signals on the basis of inter prediction comprises the steps of: inducing a motion vector predictor using motion data of a neighboring block of the current block; parsing layer data indicating the current layer to which a motion vector difference used in inter prediction of the current block belongs, in a previously defined layer structure in which the combination of one or more horizontal and vertical components of motion vector differences are divided into multiple layers; parsing index data indicating a particular combination within the current layer; inducing a motion vector difference of the current block using the layer data and index data; and inducing the motion vector for the current block by adding the motion vector difference to the motion vector predictor.
Encoding amount estimation apparatus, encoding amount estimation method and encoding amount estimation program
A coding amount estimation device includes: a feature vector generation unit that generates a feature vector on the basis of a feature map generated by an estimation target image and at least one filter set in advance; and a coding amount evaluation unit that evaluates a coding amount of the estimation target image on the basis of the feature vector.