G06T3/067

Reducing seam artifacts in 360-degree video

Provided are systems and methods for processing 360-degree video data by obtaining a 360-degree rectangular formatted projected picture, the 360-degree rectangular formatted projected picture including a first region with a region boundary and a first region area; identifying coding tree units (CTUs) within the first region area; selectively identifying a first coding unit (CU) and a second CU from within the CTUs; determining an initial QP value for the first CU; determining an initial QP value for the second CU; identifying the first CU as a region boundary CU; identifying the second CU as a non-region boundary CU; reducing the initial QP value for the first CU to a final first QP value in response to identifying the first CU as a region boundary CU, and generating an encoded video bitstream comprising: the final QP value for the first CU, and the initial QP value for the second CU.

METHOD AND NETWORK EQUIPMENT FOR TILING A SPHERE REPRESENTING A SPHERICAL MULTIMEDIA CONTENT
20200273144 · 2020-08-27 ·

A method for tiling with a set of tiles a sphere representing a spherical multimedia content, comprises: obtaining (402) an altitude for each parallel line of the sphere comprising one or several centroids of the tiles, each tile being defined as a portion of the sphere covering a tile horizontal angular amplitude and a tile vertical angular amplitude; obtaining (403) an angular position for each centroid of the tiles arranged on the parallel lines; determining (404) transformations to be applied to a reference tile to obtain the tiles of the set of tiles, each of the transformations depending on the obtained altitude and angular position of the centroid of a corresponding tile to be obtained.

Multi-User Application System Environment Engine
20200265549 · 2020-08-20 ·

A multi-user application system environment engine has an application system that, in turn, includes a simulation engine and a virtualized software environment. The simulation engine runs on top of the virtualized software environment and includes a declaration processor, a scene tree object manager, a viewer, an editor listener, and a rendering processor, coupled to the virtualized software environment, to requisition hardware resources to cause physical manifestation of an instantiated scene tree. The viewer presents to the local user a visual representation of the instantiated scene tree in its current state.

Systems and methods for image processing

An image processing method is provided, including: obtaining image data of a cavity wall of an organ; unfolding the cavity wall; and generating an image of the unfolded cavity wall. The unfolding of the cavity wall may include: obtaining a mask and a centerline of the organ; obtaining a connected region of the mask; dividing the connected region into at least one equidistant block; determining an orientation of the equidistant block in a three-dimensional coordinate system including a first direction, a second direction and a third direction; determining an initial normal vector and an initial tangent vector of a center point of the centerline; assigning a projection of the initial normal vector to a normal vector of a light direction of the center point; assigning the third direction or an reverse direction of the third direction to a tangent vector of the light direction of the center point.

HANDLING FACE DISCONTINUITIES IN 360-DEGREE VIDEO CODING
20200260120 · 2020-08-13 · ·

Systems, methods, and instrumentalities may be provided for discounting reconstructed samples and/or coding information from spatial neighbors across face discontinuities. Whether a current block is located at a face discontinuity may be determined. The face discontinuity may be a face boundary between two or more adjoining blocks that are not spherical neighbors. The coding availability of a neighboring block of the current block may be determined, e.g., based on whether the neighboring block is on the same side of the face discontinuity as the current block. For example, the neighboring block may be determined to be available for decoding the current block if it is on the same side of the face discontinuity as the current block, and unavailable if it Is not on the same side of the face discontinuity. The neighboring block may be a spatial neighboring block or a temporal neighboring block.

3D FACE IDENTITY AUTHENTICATION METHOD AND APPARATUS
20200250403 · 2020-08-06 ·

The present application provides an identity authentication method and an apparatus. The method may include obtaining a sequence of depth images containing a target face and a sequence of original two-dimensional (2D) images containing the target face, and performing identity authentication. The identity authentication may be conducted by: calculating a target face three-dimensional (3D) texture image according to the depth images containing the target face and the original 2D images containing the target face; projecting the target face 3D texture image to a 2D plane to obtain a target face 2D image; extracting feature information from the target face 2D image; comparing the feature information of the target face 2D image with feature information of a reference face 2D image to determine a similarity value; and in response to that the similarity value exceeds a first threshold, determining that the identity authentication succeeds.

METHOD & APPARATUS FOR REDUCING COLOR LEAKAGE ARTEFACTS DURING POINT CLOUD COLOR PROCESSING
20200250868 · 2020-08-06 · ·

A method for reducing color leaking artefacts in an image formed by projection processing from a 3D point cloud comprises: receiving an input image comprising the 3D point cloud; classifying the cloud into a plurality of surface patches; projecting the patches onto a plane to form a first 2D image; processing the first 2D image, by coding, transmitting and decoding, to form a final 2D image; and providing the final 2D image as an output. Processing includes independent patch processing to reduce inter-patch color leakage in the final 2D image, the independent patch processing including chroma sub-sampling pixels within each of the projected patches in the first 2D image separately; recombining the chroma sub-sampled patches to form a second 2D image; and compressing the second 2D image.

Method and system for preparing text images for optical-character recognition

The current document is directed to methods and systems that acquire an image containing text with curved text lines to generate a corresponding corrected image in which the text lines are straightened and have a rectilinear organization. The method may include identifying a page sub-image within the text-containing image, generating a text-line-curvature model for the page sub-image that associates inclination angles with pixels in the page sub-image, generating local displacements, using the text-line-curvature model, for pixels in the page sub-image, and transferring pixels from the page sub-image to a corrected page-sub-image using the local displacements to construct a corrected page sub-image in which the text lines are straightened and in which the text characters and symbols have a rectilinear arrangement.

EXTRACTION OF STANDARDIZED IMAGES FROM A SINGLE VIEW OR MULTI-VIEW CAPTURE

According to various embodiments, component information may be identified for each input image of an object. The component information may indicate a portion of the input image in which a particular component of the object is depicted. A viewpoint may be determined for each input image that indicates a camera pose for the input image relative to the object. A three-dimensional skeleton of the object may be determined based on the viewpoints and the component information. A multi-view panel corresponding to the designated component of the object that is navigable in three dimensions and that the portions of the input images in which the designated component of the object is depicted may be stored on a storage device.

AUTOMATIC VIEW MAPPING FOR SINGLE-IMAGE AND MULTI-VIEW CAPTURES

A three-dimensional (3D) skeleton may be determined based on a plurality of vertices and a plurality of faces in a two-dimensional (2D) mesh in a top-down image of an object. A correspondence mapping between a designated perspective view image and the top-down object image may be determined based on the 3D skeleton. The correspondence mapping may link a respective first location in the top-down object image to a respective second location in the designated perspective view image for each of a plurality of points in the designated perspective view image. A top-down mapped image of the object may be created by determining a first respective pixel value for each of the first locations, with each first respective pixel value being determined based on a second respective pixel value for the respective second location linked with the respective first location via the correspondence mapping.