H04N13/261

Systems, methods and computer program products for automatically extracting information from a flowchart image
11704922 · 2023-07-18 · ·

A method of extracting information from a flowchart image comprising a plurality of closed-shaped data nodes having text enclosed within, connecting lines connecting the plurality of closed-shaped data nodes and free text adjacent to the connecting lines includes receiving the flowchart image, detecting the closed-shaped data nodes, localizing the text enclosed within the closed-shaped data nodes, and masking the localized text.to generate an annotated image. Lines in the annotated image are the detected to reconstruct them as closed-shaped data nodes and connecting lines. A tree frame with the plurality of closed-shaped data nodes and the connecting lines is extracted. The free text is then localized. Chunks of the free text oriented and positioned proximally together are assembled into text blocks using an orientation-based two-dimensional clustering.

Systems, methods and computer program products for automatically extracting information from a flowchart image
11704922 · 2023-07-18 · ·

A method of extracting information from a flowchart image comprising a plurality of closed-shaped data nodes having text enclosed within, connecting lines connecting the plurality of closed-shaped data nodes and free text adjacent to the connecting lines includes receiving the flowchart image, detecting the closed-shaped data nodes, localizing the text enclosed within the closed-shaped data nodes, and masking the localized text.to generate an annotated image. Lines in the annotated image are the detected to reconstruct them as closed-shaped data nodes and connecting lines. A tree frame with the plurality of closed-shaped data nodes and the connecting lines is extracted. The free text is then localized. Chunks of the free text oriented and positioned proximally together are assembled into text blocks using an orientation-based two-dimensional clustering.

A METHOD AND APPARATUS FOR ENCODING AND DECODING VOLUMETRIC VIDEO

Methods, devices and stream are disclosed to encode and decode a volumetric content. At the encoding, the space of the volumetric content is divided in distinct sectors according to at least two different sectorizations. One atlas is generated for each sectorization or a single atlas is generated encoding all the sectorizations. At the decoding, a sectorization is selected according to the current direction and field of view, according to user's gaze navigation and according to prediction of the upcoming pose of the virtual camera controlled by the user. Sectors are selected according the selected sectorization and the current direction and field of view and only patches encoded in regions of the atlas associated with these sectors are accessed to generate the viewport image representative of the content seen from the current point of view.

Detection and ranging based on a single monoscopic frame
11595634 · 2023-02-28 · ·

One or more stereoscopic images are generated based on a single monoscopic image that may be obtained from a camera sensor. Each stereoscopic image includes a first digital image and a second digital image that, when viewed using any suitable stereoscopic viewing technique, result in a user or software program receiving a three-dimensional effect with respect to the elements included in the stereoscopic images. The monoscopic image may depict a geographic setting of a particular geographic location and the resulting stereoscopic image may provide a three-dimensional (3D) rendering of the geographic setting. Use of the stereoscopic image helps a system obtain more accurate detection and ranging capabilities. The stereoscopic image may be any configuration of the first digital image (monoscopic) and the second digital image (monoscopic) that together may generate a 3D effect as perceived by a viewer or software program.

Detection and ranging based on a single monoscopic frame
11595634 · 2023-02-28 · ·

One or more stereoscopic images are generated based on a single monoscopic image that may be obtained from a camera sensor. Each stereoscopic image includes a first digital image and a second digital image that, when viewed using any suitable stereoscopic viewing technique, result in a user or software program receiving a three-dimensional effect with respect to the elements included in the stereoscopic images. The monoscopic image may depict a geographic setting of a particular geographic location and the resulting stereoscopic image may provide a three-dimensional (3D) rendering of the geographic setting. Use of the stereoscopic image helps a system obtain more accurate detection and ranging capabilities. The stereoscopic image may be any configuration of the first digital image (monoscopic) and the second digital image (monoscopic) that together may generate a 3D effect as perceived by a viewer or software program.

VIRTUAL REALITY ENVIRONMENT
20220368878 · 2022-11-17 · ·

A three-dimensional virtual reality environment.

Method and device for converting 2D image into 3D image and 3D imaging system
11589030 · 2023-02-21 · ·

The present disclosure discloses a method and a device for converting two-dimensional (2D) images into three-dimensional (3D) images and a 3D imaging system, wherein the method comprises the following steps: acquiring 2D image to be processed; performing perspective transformation on the 2D image to obtain a left-eye image and a right-eye image respectively; adjusting a distance between the left-eye image and the right-eye image according to the result of perspective transformation; and synthesizing the left-eye image and the right-eye image after the distance adjustment. In embodiments of the present disclosure, binocular parallax images are created by performing perspective transformation on the 2D image to be processed; the distance between the left-eye image and the right-eye image after perspective transformation is adjusted to form binocular parallax and create a convergence angle, so that the images observed by naked eyes are located at different depths, thus different stereoscopic effects may be seen. The image transformation is performed on the 2D image without involving the resolution and definition of the image, so that the image quality of the 3D imaged image is the same as that of the original 2D image and the 3D imaging effect is not affected.

Method and device for converting 2D image into 3D image and 3D imaging system
11589030 · 2023-02-21 · ·

The present disclosure discloses a method and a device for converting two-dimensional (2D) images into three-dimensional (3D) images and a 3D imaging system, wherein the method comprises the following steps: acquiring 2D image to be processed; performing perspective transformation on the 2D image to obtain a left-eye image and a right-eye image respectively; adjusting a distance between the left-eye image and the right-eye image according to the result of perspective transformation; and synthesizing the left-eye image and the right-eye image after the distance adjustment. In embodiments of the present disclosure, binocular parallax images are created by performing perspective transformation on the 2D image to be processed; the distance between the left-eye image and the right-eye image after perspective transformation is adjusted to form binocular parallax and create a convergence angle, so that the images observed by naked eyes are located at different depths, thus different stereoscopic effects may be seen. The image transformation is performed on the 2D image without involving the resolution and definition of the image, so that the image quality of the 3D imaged image is the same as that of the original 2D image and the 3D imaging effect is not affected.

Methods, systems, and media for generating and rendering immersive video content
11589027 · 2023-02-21 · ·

Methods, systems, and media for generating and rendering immersive video content are provided. In some embodiments, the method comprises: receiving information indicating positions of cameras in a plurality of cameras; generating a mesh on which video content is to be projected based on the positions of the cameras in the plurality of cameras, wherein the mesh is comprised of a portion of a faceted cylinder, and wherein the faceted cylinder has a plurality of facets each corresponding to a projection from a camera in the plurality of cameras; receiving video content corresponding to the plurality of cameras; and transmitting the video content and the generated mesh to a user device in response to receiving a request for the video content from the user device.

VIRTUAL REALITY ENVIRONMENT BASED MANIPULATION OF MULTI-LAYERED MULTI-VIEW INTERACTIVE DIGITAL MEDIA REPRESENTATIONS

Various embodiments of the present disclosure relate generally to systems and methods for generating multi-view interactive digital media representations in a virtual reality environment. According to particular embodiments, a plurality of images is fused into a first content model and a first context model, both of which include multi-view interactive digital media representations of objects. Next, a virtual reality environment is generated using the first content model and the first context model. The virtual reality environment includes a first layer and a second layer. The user can navigate through and within the virtual reality environment to switch between multiple viewpoints of the content model via corresponding physical movements. The first layer includes the first content model and the second layer includes a second content model and wherein selection of the first layer provides access to the second layer with the second content model.