H04N19/543

Device and method for recognizing motion in vehicle

A device for recognizing a motion in a vehicle according to an embodiment of the present disclosure may include a camera for acquiring a user image, and a controller that divides the user image into a first region, a second region, and a third region in which the first region and the second region overlap each other, and recognizes a motion of a user occurring in at least one of the first region, the second region, or the third region.

Encoder, decoder, encoding method, and decoding method

Provided is an encoder that achieves further improvement. The encoder includes processing circuitry and memory. Using the memory, the processing circuitry: obtains two prediction images from two reference pictures; derives a luminance gradient value of each pixel position in each of the two prediction images; derives a luminance local motion estimation value of each pixel position in a current block; generates a luminance final prediction image using a luminance value and the luminance gradient value in each of the two prediction images, and the luminance local motion estimation value of the current block; and generates a chrominance final prediction image using at least one of the luminance gradient value of each of the two prediction images or the luminance local motion estimation value of the current block, and chrominance of each of the two prediction images.

Simplification of hash-based motion searching

Methods, systems and device for hash-based motion estimation in video coding are described. An exemplary method of video processing includes determining, for a conversion between a current block of a video and a bitstream representation of the video, motion information associated with the current block using a hash-based motion search, a size of the current block being M×N, M and N being positive integers and M being not equal to N, applying, based on the motion information and a video picture comprising the current block, a prediction for the current block, and performing, based on the prediction, the conversion.

Simplification of hash-based motion searching

Methods, systems and device for hash-based motion estimation in video coding are described. An exemplary method of video processing includes determining, for a conversion between a current block of a video and a bitstream representation of the video, motion information associated with the current block using a hash-based motion search, a size of the current block being M×N, M and N being positive integers and M being not equal to N, applying, based on the motion information and a video picture comprising the current block, a prediction for the current block, and performing, based on the prediction, the conversion.

GEOMETRIC PARTITION MODE WITH INTRA BLOCK COPY
20230034458 · 2023-02-02 · ·

In response to a current block being partitioned into a first partition and a second partition based on a geometric partition mode, a first prediction type for the first partition and a second prediction type for the second partition are determined based on a flag associated with one of intra prediction, inter prediction, and intra block copy (IBC). Each of the first and second prediction types is one of the intra prediction, the inter prediction, and the IBC. Based on the first prediction type, a first prediction mode for the first partition of the current block is determined. Based on the second prediction type, a second prediction mode for the second partition of the current block is determined. The first partition of the current block is reconstructed based on the first prediction mode and the second partition of the current block is reconstructed based on the second prediction mode.

Interaction between core transform and secondary transform
11616966 · 2023-03-28 · ·

A video decoder receives data from a bitstream for a block of pixels to be decoded as a current block of a current picture of a video. The video decoder receives from the bitstream first and second signaled indices for the current block. The video decoder determines first and second merge indices from the first and second signaled indices. The video decoder uses the first and second merge indices to select first and second motion candidates, respectively. The video decoder generates a set of prediction samples in ALWIP mode and performs an inverse secondary transform and an inverse primary transform to generate a set of residual samples of the current block. Enabling or selection of secondary transform and/or primary transform depends on size, width, and/or height for the current block. The video decoder reconstructs the current block by using the set of residual samples and the set of prediction samples.

Image encoding/decoding method and apparatus involving merge candidate list and triangular shape partitions
11490077 · 2022-11-01 · ·

An image encoding/decoding apparatus according to the present invention can configure a merge candidate list of a current block to which a diagonal motion partition is applied, derive the motion information of the current block on the basis of the merge candidate list and a merge candidate index, and perform inter-prediction on the current block on the basis of the derived motion information.

Motion compensation method and module, chip, electronic device and storage media

The present disclosure relates to a motion compensation method and module, a chip, an electronic device, and a storage medium, to improve the problem of haloes easily appearing on the edges of moving objects.

Motion compensation method and module, chip, electronic device and storage media

The present disclosure relates to a motion compensation method and module, a chip, an electronic device, and a storage medium, to improve the problem of haloes easily appearing on the edges of moving objects.

BIDIRECTIONAL OPTICAL FLOW BASED VIDEO CODING AND DECODING
20220329842 · 2022-10-13 ·

Devices, systems and methods for sample refinement and filtering method for video coding are described. In an exemplary aspect, a method for video processing includes modifying for a conversion between a block of a video and a bitstream representation of the video, a refinement value for a prediction sample in the block by applying a clipping operation to refinement value. The refinement value is derived based on a gradient value of an optical flow coding process. An output of the clipping operation is within a range. The method also includes refining the prediction sample based on the refinement value and performing the conversion based on the refined prediction sample.