METHOD AND SYSTEM FOR SYNTHESIZING NOVEL VIEW IMAGE ON BASIS OF MULTIPLE 360 IMAGES FOR 6-DEGREES OF FREEDOM VIRTUAL REALITY
20220358712 · 2022-11-10
Assignee
Inventors
Cpc classification
International classification
Abstract
A method and a system for synthesizing a novel-view image based on multiple 360 images for 6DOF virtual reality, in which a large-scale 6-DOF virtual environment is implemented, and a scene is synthesized at a novel viewpoint, are provided. The method includes performing a 3D reconfiguration procedure for the 360 images to recover 3D geometric information, and to reconfigure a virtual data map in which the multiple 360 images are integrated into one image, producing a view image corresponding to a viewpoint of a user by applying a view synthesis algorithm of projection & vertex warping process using a reference image which is closest to a viewpoint extracted from the virtual data map, and blending view images for 6DoF through a section formula for inner split based on a distance between a position of the reference image and a position of the viewpoint.
Claims
1. A method for synthesizing a novel-view image based on multiple 360 images for 6DOF virtual reality, the method comprising: performing a 3D reconfiguration procedure for the 360 images to recover 3D geometric information, and to reconfigure a virtual data map in which the multiple 360 images are integrated into one image; producing a view image corresponding to a viewpoint of a user by applying a view synthesis algorithm of projection & vertex warping process using a reference image which is closest to a viewpoint extracted from the virtual data map; and blending view images for 6DoF through a section formula for inner split based on a distance between a position of the reference image and a position of the viewpoint.
2. The method of claim 1, wherein the reconfiguring of the virtual data map includes: reconfiguring the virtual data map, which is a reference sphere of a sphere mesh having a triangular surface and a plurality of vertexes by subdividing a Icosahedron sphere several times, with respect to the 360 images.
3. The method of claim 2, wherein the reconfiguring of the virtual data map includes: performing the 3D reconfiguration in a structure from motion (SfM) scheme for view synthesis based on the multiple 360 images; and reconfiguring the virtual data map in which the 3D geometric information is integrated into one image.
4. The method of claim 3, wherein the 3D geometric information includes: a point cloud for the 360 images, a 3D mesh based on the point cloud, and an external parameter group (a camera location) of a camera, which indicates a posture of the camera.
5. The method of claim 2, wherein the producing of the view image includes: acquiring the viewpoint corresponding to each vertex of the virtual data map by using the reference image mapped to the reference sphere, and acquiring a position, in which the vertex is moved, as the viewpoint is projected to a novel sphere; inducing movement of pixels by moving vertexes of the reference sphere to the acquired position; and producing the view image corresponding to a field-of view at a user viewpoint by positioning a camera into the novel sphere.
6. The method of claim 5, wherein the acquiring of the viewpoint and the position includes: using a single reference image closest to the viewpoint or two reference images closes to the viewpoint.
7. The method of claim 5, wherein the blending of the view image includes: employing a weighted blending scheme for blending pixels inversely proportional to the distance, with respect to at least two reference images, to prevent an environment of the reference image from being switched.
8. The method of claim 7, wherein the blending of the view image includes: calculating pixel values of the view image at the final stage through the section formula, after acquiring a distance between the novel sphere and the reference sphere to perform the weighted blending scheme for the pixels.
9. The method of claim 8, wherein the section formula is an equation for acquiring weighted blending.
10. A computer program stored in a computer-readable medium to execute a method for synthesizing a novel-view image based on multiple 360 images for 6DOF virtual reality, wherein the method for synthesizing the novel-view image based on the multiple 360 images for the 6DOF virtual reality, includes: performing a 3D reconfiguration procedure for the 360 images to recover 3D geometric information, and to reconfigure a virtual data map in which the multiple 360 images are integrated into one image; producing a view image corresponding to a viewpoint of a user by applying a view synthesis algorithm of projection & vertex warping process using a reference image which is closest to a viewpoint extracted from the virtual data map; and blending view images for 6DoF through a section formula for inner split based on a distance between a position of the reference image and a position of the viewpoint.
11. A system for synthesizing a novel-view image based on multiple 360 images for 6DOF virtual reality, the system comprising: a reconfiguring unit configured to perform a 3D reconfiguration procedure for the 360 images to recover 3D geometric information, and to reconfigure a virtual data map in which the multiple 360 images are integrated into one image; a processing unit configured to generate a view image corresponding to a viewpoint of a user by applying a view synthesis algorithm of projection & vertex warping process using a reference image which is closest to a viewpoint extracted from the virtual data map; and a blending unit configured to blend view images for 6DoF through a section formula for inner split based on a distance between a position of the reference image and a position of the viewpoint.
12. The system of claim 11, wherein the reconfiguring unit reconfigures the virtual map, which is a reference sphere of a sphere mesh having a triangular surface and a plurality of vertexes by subdividing a Icosahedron sphere several times, with respect to the 360 images.
13. The system of claim 12, wherein the reconfiguring unit performs the 3D reconfiguration in a structure from motion (SfM) scheme for view synthesis based on the multiple 360 images, and reconfigures the virtual data map in which the 3D geometric information is integrated into one image.
14. The system of claim 12, wherein the processing unit includes: an acquiring unit configured to acquire the viewpoint corresponding to each vertex of the virtual data map by using the reference image mapped to the reference sphere, and acquire a position, in which the vertex is moved, as the viewpoint is projected to a novel sphere; a pixel unit configured to induce movement of pixels by moving vertexes of the reference sphere to the acquired position; and a generator configured to generate the view image corresponding to a field-of view at a user viewpoint by positioning a camera into the novel sphere.
15. The system of claim 14, wherein the acquiring unit uses a single reference image closest to the viewpoint or two reference images closes to the viewpoint.
16. The system of claim 14, wherein the blending unit employs a weighted blending scheme for blending pixels inversely proportional to the distance, with respect to at least two reference images, to prevent an environment of the reference image from being switched.
17. The system of claim 16, wherein the blending unit calculates pixel values of the view image at the final stage through the section formula, after acquiring a distance between the novel sphere and the reference sphere to perform the weighted blending scheme for the pixels.
Description
DESCRIPTION OF THE DRAWINGS
[0031]
[0032]
[0033]
[0034]
[0035]
[0036]
[0037]
[0038]
DESCRIPTION
[0039] Hereinafter, exemplary embodiments of the present disclosure will be described in detail with reference to the accompanying drawings. However, the present disclosure is not limited or restricted by the embodiments. Further, the same reference signs/numerals in the drawings indicate the same members.
[0040] Furthermore, the terminologies used herein are used to properly express the embodiments of the present disclosure, and may be changed according to the intent of a viewer or an operator in the field to which the present disclosure pertains. Accordingly, definition of the terms should be made according to the overall disclosure set forth herein.
[0041] The subject matter of the present disclosure is to provide a method for synthesizing a user view image in real time, enabling the user to experience a 6DoF in a wider space.
[0042] In particular, the subject matter of the present disclosure is to construct a large-scale 6DOF virtual environment by using multiple 360 images, and to synthesize scenes at a novel viewpoint. When a user view is synthesized based on a single 360 image in real time, a user may freely experience a view through a complete head motion of a 6DoF, but a space, in which the user is movable, is confined within an image context. Therefore, the present disclosure suggests a process of minimizing an error of a synthesizing result by considering the multiple 360 images and by performing weighted blending, as the virtual data map based on reality in a wider space is reconfigured.
[0043] Hereinafter, a method for synthesizing a novel-view image based on multiple 360 images for 6DOF virtual reality allowing the experience of a virtual environment in a wider area and providing smooth switching between 360 images, and a system for the same will be described in detail with reference to
[0044]
[0045] Hereinafter, a method for synthesizing a novel-view image based on multiple 360 images for 6DOF virtual reality will be described with reference to
[0046] Referring to
[0047] In step 110 and step 210, the 3D reconfiguration procedure, such as a structure from motion (SfM) procedure serving as a pre-processing procedure may be performed, to recover the 3D geometric information of scenes, which are necessary for view-synthesis, from the multiple 360 images. In this case, the estimated 3D geometric information may include a point cloud for 360 images, a 3D mesh based on the point cloud, and an external parameter group (a camera location) of a camera, which indicates the posture of the camera.
[0048] Thereafter, step 110 is to reconfigure a virtual data map in which all 3D geometric information is integrated into one.
[0049] Referring to
[0050] Accordingly, in step S110, the virtual map is reconfigured on a reference sphere (a sphere positioned at the rightmost part of
[0051] In step 120 and step 220, a view image corresponding to the viewpoint of the user is generated by applying a view synthesis algorithm of the projection & vertex warping process by using a reference image closest to a viewpoint extracted from the virtual data map.
[0052] In step 120, a novel view synthesis algorithm may be employed, based on at least one reference image after acquiring the virtual data map of the schemes. The reference image closet to the novel viewpoint should have a view the most view to a novel view. In this regard, in step 120, a method of employing the closest single reference image is described.
[0053] In more detail, step 120 may include the step (not illustrated) of acquiring a viewpoint corresponding to each vertex of the virtual data map by using the reference image mapped to the reference sphere, and acquiring a position, in which the vertex is moved, as the viewpoint is projected to the novel sphere, the step (not illustrated) of inducing the movement of pixels by moving the vertexes of the reference sphere to the acquired position, and the step of producing a view image corresponding to a field-of view at a user viewpoint by positioning the camera into the novel sphere.
[0054] In this case, in the step of acquiring the viewpoint and the position, a single reference image closest to the viewpoint or two reference images closes to the viewpoint may be used.
[0055] For example, referring to
[0056] Accordingly, in step 120, the viewpoint is projected to a novel sphere, the position, to which the vertex is moved, is acquired, and the vertexes of the reference images are moved to the relevant position, thereby inducing the movement of the pixels. Thereafter, in step 120, the novel view image may be generated through the sphere matched to the field of view of the display.
[0057] For another example, regarding the description, which is made with reference to
[0058] In step 130, view images for 6DoF are blended through a section formula for inner split based on the distance between the position of the reference image and the position of the viewpoint.
[0059] In step 130, at least two reference images are considered, and a weighted blending scheme for blending pixels inversely proportional to the distance may be used to prevent a screen image, which is generated when the reference image is changed, from being switched.
[0060] In addition, in step 130, as illustrated in
[0061] In this case, the section formula, which is Equation 1, may be an equation for weighted blending.
[0062] In Equation 1, d.sub.1 and d.sub.2 denote the distances between the novel sphere and the reference spheres, and Pixel Interpolated denote pixel values of novel view images generated based the reference sphere.
[0063]
[0064] In more detail,
[0065] In addition,
[0066] In Experiment 1 (Exp.1), the total of 22 images are synthesized at 22 points positioned between positions of the two reference images spaced apart from each other by 50 m. Thereafter, the synthesized images were compared with real-measured images through both a peak signal-to-noise ratio (PSNR) measuring scheme and a structural similarity metric (SSIM) measuring scheme.
[0067] In Experiment 2 (Exp.2), the total of 44 images and frames were synthesized at uniform distances between three reference images spaced apart from each other by 1.75 m. Next, the smoothness of all images (reference and synthesis) was evaluated by comparing a previous frame with each frame.
[0068]
[0069] In more detail,
[0070] In
[0071] In detail, in the case of M1, it may be recognized that the image quality of the synthesized image is gradually reduced, as the distance to the reference image is increased up to 25 m, and, as the reference image is changed after the distance of 25 m, the image quality of the synthesized image is gradually recovered toward the changed reference image. To the contrary, in the case of M2, it may be recognized that the image quality is maintained, even though two reference images are interpolated and a novel view is away from the reference images,
[0072] Referring to
[0073] Therefore, in the method for synthesizing a novel-view image based on multiple 360 images for 6DoF virtual reality according to an embodiment of the present disclosure, the projection & warping scheme is utilized for a sphere by using the reconfigured 3D scenes, a desired view is synthesized based on a single 360 image, and the weighted blending is applied to the plurality of reference views to interpolate a plurality of synthesized views, thereby prospecting the growth of VR content based on 360 images through a 6DoF VR system according to the present disclosure.
[0074]
[0075] Referring to
[0076] To this end, a system 900 for synthesizing a novel-view image based on multiple 360 images for 6DOF virtual reality includes a reconfiguration unit 910, a processing unit 920, and a blending unit 930
[0077] The reconfiguring unit 910 reconfigures 360 images to three-dimension (3D) images to recover 3D geometric information, and reconfigures a virtual data map in which the multiple 360 images are integrated into one image.
[0078] The reconfiguration unit 910 may perform the 3D reconfiguration procedure, such as a structure from motion (SfM) procedure serving as a pre-processing procedure, to recover the 3D geometric information of scenes, which are necessary for view-synthesis, from the multiple 360 images. In this case, the estimated 3D geometric information may include a point cloud for 360 images, a 3D mesh based on the point cloud, and an external parameter group (a camera location) of a camera, which indicates the posture of the camera.
[0079] The reconfiguration unit 910 may reconfigure a virtual data map in which all 3D geometric information is unified.
[0080] According to an embodiment of the present disclosure, the reconfiguration unit 910 of the system 900 for synthesizing a novel-view image based on multiple 360 images for 6DOF virtual reality may reconfigure a sphere mesh having a triangular surface and a plurality of vertexes by sub-dividing a Icosahedron sphere several times, in association with the 360 images. The sphere mesh having numerous vertexes is used to express the moving of a pixel in the vertex warping process.
[0081] The processing unit 920 generates a view image corresponding to the viewpoint of the user by applying a view synthesis algorithm of the projection & vertex warping process by using a reference image closes to a viewpoint extracted from the virtual data map.
[0082] The processing unit 920 may employ a novel view synthesis algorithm, based on at least one reference image after acquiring the virtual data map of the schemes. The reference image closet to the novel viewpoint should have a view the most view to a novel view. In this regard, a method in which the processing unit 920 uses the closest single reference image will be described below.
[0083] In more detail, the processing unit 920 may include an acquisition unit (not illustrated) to acquire a viewpoint corresponding to each vertex of the virtual data map by using the reference image mapped to the reference sphere, and acquire a position, in which the vertex is moved, as the viewpoint is projected to the novel sphere, a pixel unit (not illustrated) to induce the movement of pixels by moving the vertexes of the reference sphere to the acquired position, and a generator (not illustrated) to generate a view image corresponding to a field-of view at a user viewpoint by positioning the camera into the novel sphere.
[0084] In this case, the acquiring unit may use a single reference image closes to the viewpoint or two reference images closes to the viewpoint.
[0085] For example, referring to
[0086] Accordingly, the processing unit 920 may project the viewpoint to a novel sphere, acquire the position, to which the vertex is moved, and move the vertexes of the reference images to the relevant position, thereby inducing the movement of the pixels. Thereafter, the processing unit 920 may generate the novel view image through the sphere matched to the field of view of the display.
[0087] For another example, the processing unit 920, which performs the projection & vertex warping by using two reference images, may acquire viewpoints in a three-dimension (3D) model corresponding to a vertex of reference image 1 and a vertex of reference image 2, and project the viewpoints to the novel sphere, thereby acquiring positions to which the vertexes are moved. Thereafter, the processing unit 920 may induce the movement of pixels by applying 360 images to the positions, that is, moving the vertexes of reference image 1 and reference image 2. Accordingly, at the final image, the processing unit 920 may position the camera into the novel sphere, and may generate the novel view image corresponding to the field-of-view at the user viewpoints.
[0088] The blending unit 930 blends view images for 6DoF through a section formula for inner split based on the distance between the position of the reference image and the position of the viewpoint.
[0089] The blending unit 930 may consider at least two reference images, and use a weighted blending scheme for blending pixels inversely proportional to the distance, to prevent a screen image, which is generated when the reference image is changed, from being switched.
[0090] In addition, the blending unit 930 may acquire the distance between the novel sphere and reference sphere 1, and the distance between the novel sphere and reference 2, and then may calculate the pixel values of the view image at the final stage through the section formula to perform the weighted blending for the pixels. In this case, the section formula, which is Equation 1 as described above, may be an equation for weighted blending.
[0091] It is obvious to those skilled in the art that the system according to the present disclosure includes all features described with reference to
[0092] The foregoing devices may be realized by hardware elements, software elements and/or combinations thereof. For example, the devices and components illustrated in the exemplary embodiments of the inventive concept may be implemented in one or more general-use computers or special-purpose computers, such as a processor, a controller, an arithmetic logic unit (ALU), a digital signal processor, a microcomputer, a field programmable array (FPA), a programmable logic unit (PLU), a microprocessor or any device which may execute instructions and respond. A processing unit may perform an operating system (OS) or one or software applications running on the OS. Further, the processing unit may access, store, manipulate, process and generate data in response to execution of software. It will be understood by those skilled in the art that although a single processing unit may be illustrated for convenience of understanding, the processing unit may include a plurality of processing elements and/or a plurality of types of processing elements. For example, the processing unit may include a plurality of processors or one processor and one controller. Also, the processing unit may have a different processing configuration, such as a parallel processor.
[0093] Software may include computer programs, codes, instructions or one or more combinations thereof and configure a processing unit to operate in a desired manner or independently or collectively control the processing unit. Software and/or data may be permanently or temporarily embodied in any type of machine, component, physical equipment, virtual equipment, computer storage medium or unit or transmitted signal waves so as to be interpreted by the processing unit or to provide instructions or data to the processing unit. Software may be dispersed throughout computer systems connected over networks and be stored or executed in a dispersion manner. Software and data may be recorded in one or more computer-readable storage medium.
[0094] The method according to an embodiment may be implemented in the form of a program instruction and may be recorded in a computer-readable recording medium. The computer-readable storage medium may also include program instructions, data files, data structures, or a combination thereof. The program instructions recorded in the medium may be designed and configured specially for the embodiment or may be known and available to those skilled in computer software. The computer-readable storage medium may include a hardware device, which is specially configured to store and execute program instructions, such as magnetic media (e.g., a hard disk drive and a magnetic tape), optical media (e.g., CD-ROM and DVD), magneto-optical media (e.g., a floptical disk), a read only memory (ROM), a random access memory (RAM), or a flash memory. Examples of program instructions include not only machine language codes created by a compiler, but also high-level language codes that are capable of being executed by a computer by using an interpreter or the like. The described hardware devices may be configured to act as one or more software modules to perform the operations of the above-described embodiments, or vice versa.
[0095] While embodiments have been shown and described with reference to the accompanying drawings, it will be apparent to those skilled in the art that various modifications and variations can be made from the foregoing descriptions. For example, adequate effects may be achieved even if the foregoing processes and methods are carried out in different order than described above, and/or the aforementioned elements, such as systems, structures, devices, or circuits, are combined or coupled in different forms and modes than as described above or be substituted or switched with other components or equivalents.
[0096] Therefore, other implements, other embodiments, and equivalents to claims are within the scope of the following claims.