Patent classifications
G06T3/16
ENCODING METHOD, PLAYING METHOD AND APPARATUS FOR IMAGE STABILIZATION OF PANORAMIC VIDEO, AND METHOD FOR EVALUATING IMAGE STABILIZATION ALGORITHM
An encoding method, playing method and apparatus for image stabilization of panoramic video, and a method for evaluating image stabilization algorithm are provided. The image stabilization method for the panoramic video is applicable to an electronic apparatus including a processor. In the method, a plurality of image frames of a panoramic video is captured, and each image frame is transformed into a plurality of projection frames on a plurality of faces of a cubemap. Then, variations of triaxial displacements and attitude angles between the projection frames transformed onto each of the faces and adjacent in time are calculated. The variations of triaxial displacements and attitude angles are smoothed and recorded as movement information. While playing the panoramic video, the panoramic video is corrected by the movement information and played. Thus, it is possible to reduce the amount of calculation required for the stabilization calculations on the captured video.
Interconnected 360 video virtual travel
System and method for synchronizing display of panoramic videos is disclosed. In one embodiment, a method of synchronizing display of panoramic video frames is discloses that includes transmitting a first panoramic image, displaying one or more first perspective frames based on the first panoramic image, determining an alignment parameter in response to a selected orientation of the first panoramic image, wherein the selected orientation has an associated one of the one or more first perspective frames, transmitting a second panoramic image, and displaying one or more second perspective frames based on the second panoramic image, wherein displaying the one or more second perspective frames includes displaying an initial one of the one or more second perspective frames having an orientation based on the alignment parameter.
4D convolutional neural networks for video recognition
This disclosure includes technologies for video recognition in general. The disclosed system can automatically detect various types of actions in a video, including reportable actions that cause shrinkage in a practical application for loss prevention in the retail industry. The temporal evolution of spatio-temporal features in the video are used for action recognition. Such features may be learned via a 4D convolutional operation, which is adapted to model low-level features based on a residual 4D block. Further, appropriate responses may be invoked if a reportable action is recognized.
SPHERICAL VIDEO EDITING
Systems and methods provide for editing of spherical video data. In one example, a computing device can receive a spherical video (or a video associated with an angular field of view greater than an angular field of view associated with a display screen of the computing device), such as by a built-in spherical video capturing system or acquiring the video data from another device. The computing device can display the spherical video data. While the spherical video data is displayed, the computing device can track the movement of an object (e.g., the computing device, a user, a real or virtual object represented in the spherical video data, etc.) to change the position of the viewport into the spherical video. The computing device can generate a new video from the new positions of the viewport.
COORDINATE MAPPING FOR RENDERING PANORAMIC SCENE
A projection metadata system and method are provided for providing projection metadata to a rendering system for use in rendering a panoramic scene. A rendering system and method are provided for rendering the panoramic scene using the projection metadata. The projection metadata may comprise machine readable data which may directly provide the coordinate mapping to be used in by the rendering system to render the panoramic scene. For example, an executable script may be provided which, when executed, carries out the coordinate mapping. Compared to known ways of signalling a rendering system which projection to use, which may conventionally involve selecting a projection from a number of predefined projections and signalling the selection, the use of projection metadata may provide a high degree of flexibility, as different projections may be defined for, e.g., different spatial and/or temporal parts of the panoramic scene, different rendering techniques, etc.
Method and apparatus for generating and encoding projection-based frame with 360-degree content represented by base projection face and lateral projection faces packed in pyramid projection layout
A video processing method includes receiving an omnidirectional content corresponding to a sphere, generating a projection-based frame according to the omnidirectional content and a pyramid projection layout, and encoding, by a video encoder, the projection-based frame to generate a part of a bitstream. The projection-based frame has a 360-degree content represented by a base projection face and a plurality of lateral projection faces packed in the pyramid projection layout. The base projection face and the lateral projection faces are obtained according to at least projection relationship between a pyramid and the sphere.
Rendering 360 depth content
As user device can receive and display 360 panoramic content in a 360 depth format. 360 depth content can comprise 360 panoramic image data and corresponding depth information. To display 360 depth content, the user device can generate a 3D environment based on the 360 depth content and the current user viewpoint. A content display module on the user device can render 360 depth content using a standard 3D rendering pipeline modified to render 360 depth content. The content display module can use a vertex shader or fragment shader of the 3D rendering pipeline to interpret the depth information of the 360 depth content into the 3D environment as it is rendered.
IMAGE PROCESSING APPARATUS AND METHOD USING VIDEO SIGNAL OF PLANAR COORDINATE SYSTEM AND SPHERICAL COORDINATE SYSTEM
The present disclosure relates to an image processing apparatus and method for transforming image data of a spherical coordinate system corresponding to image data of a planar coordinate system without transforming the image data of the planar coordinate system, and stitching the transformed image data of the spherical coordinate system to the image data of the planar coordinate system.
HIGHLY SCALABLE CLUSTER ENGINE FOR HOSTING SIMULATIONS OF OBJECTS INTERACTING WITHIN A SPACE
A system for simulating objects in a three dimensional virtual space, comprising a host computing node coupled to a network, padding spheres defined in the virtual space, one padding sphere fully encasing each cube in an octree; a plurality of objects, each object tracked for position in the virtual space. A visibility sphere associated with each object is provided, and at least one display, displaying objects in the virtual space. One of the objects becoming a target object, a search determines objects to be rendered in the display, identifying all padding spheres overlapped at least partially by the visibility sphere of the target object as candidate padding spheres, and identifying objects within or overlapping the candidate padding spheres, determining a visibility ratio for each object of maximum dimension of the object divided by distance from the target object, any object having a visibility to be rendered in the display.
Method and apparatus for generating three-dimensional panoramic video
A method and an apparatus for generating 3D panoramic video are provided. In the method, plural frames are captured from a panoramic video. Each frame is transformed into a polyhedral mapping projection comprising side planes, a top plane and a bottom plane. Displacements of pixels in the side planes are calculated by using the side planes of each frame, and displacements of pixels in the top plane and the bottom plane are calculated by using the displacements of the side planes. Then, the pixels in the side planes, the top plane and the bottom plane of each frame are shifted according the displacements of the polyhedral mapping projection to generate a shifted polyhedral mapping projection. The shifted polyhedral mapping projection is transformed into a shifted frame with 2D space format. The shifted frames and corresponding frames construct 3D images and the 3D images are encoded into a 3D panoramic video.