Patent classifications
G06T15/20
DYNAMIC WIDGET PLACEMENT WITHIN AN ARTIFICIAL REALITY DISPLAY
The disclosed computer-implemented method may include (1) identifying a trigger element within a field of view presented by a display element of an artificial reality device, (2) determining a position of the trigger element within the field of view, (3) selecting a position within the field of view for a virtual widget based on the position of the trigger element, and (4) presenting the virtual widget at the selected position via the display element. Various other methods, systems, and computer-readable media are also disclosed.
3D BUILDING GENERATION USING TOPOLOGY
Embodiments provide systems and methods for three-dimensional building generation from machine learning and topological models. The method uses topology models that are converted into vertices and edges. A BGAN (Building generative adversarial network) is used to create fake vertices/edges. The BGAN is then used to generate random samples from seen sample of different structures of building based on relationship of vertices and edges. The embeddings are then fed into a machine trained network to create a digital structure from the image.
Horizon-based navigation
Systems, devices, methods, and computer-readable media for horizon-based navigation. A method can include receiving image data corresponding to a geographical region in a field of view of an imaging unit and in which the device is situated, based on the received image data, generating, by the processing unit, an image horizon corresponding to a horizon of the geographical region and from a perspective of the imaging unit, projecting three-dimensional (3D) points of a 3D point set of the geographical region to an image space of the received image data resulting in a synthetic image, generating, by the processing unit, a synthetic image horizon of the synthetic image, and responsive to determining the image horizon sufficiently correlates with the synthetic image horizon, providing a location corresponding to a perspective of the synthetic image as a location of the processing unit.
Horizon-based navigation
Systems, devices, methods, and computer-readable media for horizon-based navigation. A method can include receiving image data corresponding to a geographical region in a field of view of an imaging unit and in which the device is situated, based on the received image data, generating, by the processing unit, an image horizon corresponding to a horizon of the geographical region and from a perspective of the imaging unit, projecting three-dimensional (3D) points of a 3D point set of the geographical region to an image space of the received image data resulting in a synthetic image, generating, by the processing unit, a synthetic image horizon of the synthetic image, and responsive to determining the image horizon sufficiently correlates with the synthetic image horizon, providing a location corresponding to a perspective of the synthetic image as a location of the processing unit.
Method and system for determining a current gaze direction
A method for determining a current gaze direction of a user in relation to a three-dimensional (“3D”) scene, the 3D scene sampled by a rendering function to produce a two-dimensional (“2D”) projection image of the 3D scene, the sampling performed based on a virtual camera in turn being associated with a camera position and camera direction in the 3D scene. The method includes determining, by a gaze direction detection means, a first gaze direction of the user related to the 3D scene at a first gaze time point. The method includes determining a time-dependent virtual camera 3D transformation representing a change of a virtual camera position and/or virtual camera direction between the first gaze time point and a second sampling. The method includes determining the current gaze direction as a modified gaze direction calculated based on the first gaze direction and an inverse of the time-dependent virtual camera 3D transformation.
Systems and methods for reconstruction and rendering of viewpoint-adaptive three-dimensional (3D) personas
An exemplary method includes maintaining a receiver-side mesh-vertices list, receiving duplicative-vertex information from a sender, and responsively reducing the receiver-side mesh-vertices list in accordance with the received duplicative-vertex information, and rendering, using the reduced receiver-side mesh-vertices list, viewpoint-adaptive three-dimensional (3D) personas of a subject at least in part by weighting video pixel colors from different video-camera vantage points of video cameras that capture video streams of the subject, the weighting being performed according to a respective geometric relationship of each video-camera vantage point to a user-selected viewpoint.
Systems and methods for reconstruction and rendering of viewpoint-adaptive three-dimensional (3D) personas
An exemplary method includes maintaining a receiver-side mesh-vertices list, receiving duplicative-vertex information from a sender, and responsively reducing the receiver-side mesh-vertices list in accordance with the received duplicative-vertex information, and rendering, using the reduced receiver-side mesh-vertices list, viewpoint-adaptive three-dimensional (3D) personas of a subject at least in part by weighting video pixel colors from different video-camera vantage points of video cameras that capture video streams of the subject, the weighting being performed according to a respective geometric relationship of each video-camera vantage point to a user-selected viewpoint.
Dynamic image capturing apparatus and method using arbitrary viewpoint image generation technology
Embodiments relate to a dynamic image capturing method and apparatus using an arbitrary viewpoint image generation technology, in which an image of background content displayed on a background content display unit or an image of background content implemented in a virtual space through a chroma key screen, having a view matching to a view of seeing a subject at a viewpoint of a camera is generated, and a final image including the image of the background content and a subject area is obtained.
Systems, methods, and media for displaying interactive augmented reality presentations
Systems, methods, and media for displaying interactive augmented reality presentations are provided. In some embodiments, a system comprises: a plurality of head mounted displays, a first head mounted display comprising a transparent display; and at least one processor, wherein the at least one processor is programmed to: determine that a first physical location of a plurality of physical locations in a physical environment of the head mounted display is located closest to the head mounted display; receive first content comprising a first three dimensional model; receive second content comprising a second three dimensional model; present, using the transparent display, a first view of the first three dimensional model at a first time; and present, using the transparent display, a first view of the second three dimensional model at a second time subsequent to the first time based one or more instructions received from a server.
Systems, methods, and media for displaying interactive augmented reality presentations
Systems, methods, and media for displaying interactive augmented reality presentations are provided. In some embodiments, a system comprises: a plurality of head mounted displays, a first head mounted display comprising a transparent display; and at least one processor, wherein the at least one processor is programmed to: determine that a first physical location of a plurality of physical locations in a physical environment of the head mounted display is located closest to the head mounted display; receive first content comprising a first three dimensional model; receive second content comprising a second three dimensional model; present, using the transparent display, a first view of the first three dimensional model at a first time; and present, using the transparent display, a first view of the second three dimensional model at a second time subsequent to the first time based one or more instructions received from a server.