Patent classifications
H04N2013/0085
Image processing apparatus, image processing method, and storage medium
An image processing apparatus acquires first shape information representing a three-dimensional shape about an object located within an image capturing region based on one or more images obtained by one or more imaging apparatuses for performing image capturing of the image capturing region from a plurality of directions, acquires second shape information representing a three-dimensional shape about an object located within the image capturing region based on one or more images obtained by one or more imaging apparatuses, acquires viewpoint information indicating a position and direction of a viewpoint, and generates a virtual viewpoint image corresponding to the position and direction of the viewpoint indicated by the acquired viewpoint information based on the acquired first shape information and the acquired second shape information, such that at least a part of the object corresponding to the second shape information is displayed in a translucent way within the virtual viewpoint image.
Multispectral stereo camera self-calibration algorithm based on track feature registration
The present invention discloses a multispectral stereo camera self-calibration algorithm based on track feature registration, and belongs to the field of image processing and computer vision. Optimal matching points are obtained by extracting and matching motion tracks of objects, and external parameters are corrected accordingly. Compared with an ordinary method, the present invention uses the tracks of moving objects as the features required for self-calibration. The advantage of using the tracks is good cross-modal robustness. In addition, direct matching of the tracks also saves the steps of extraction and matching the feature points, thereby achieving the advantages of simple operation and accurate results.
VIRTUAL REALITY ENVIRONMENT BASED MANIPULATION OF MULTI-LAYERED MULTI-VIEW INTERACTIVE DIGITAL MEDIA REPRESENTATIONS
Various embodiments of the present disclosure relate generally to systems and methods for generating multi-view interactive digital media representations in a virtual reality environment. According to particular embodiments, a plurality of images is fused into a first content model and a first context model, both of which include multi-view interactive digital media representations of objects. Next, a virtual reality environment is generated using the first content model and the first context model. The virtual reality environment includes a first layer and a second layer. The user can navigate through and within the virtual reality environment to switch between multiple viewpoints of the content model via corresponding physical movements. The first layer includes the first content model and the second layer includes a second content model and wherein selection of the first layer provides access to the second layer with the second content model.
Apparatus and method for generating view images
An apparatus for generating view images for a scene comprises a store (101) which stores three dimensional scene data representing the scene from a viewing region. The three dimensional scene data may e.g. be images and depth maps captured from capture positions within the viewing region. A movement processor (105) receives motion data, such as head or eye tracking data, for a user and determines an observer viewing position and an observer viewing orientation from the motion data. A change processor (109) determines an orientation change measure for the observer viewing orientation and an adapter (111) is arranged to reduce a distance from the observer viewing position relative to the viewing region in response to the orientation change measure. An image generator (103) generates view images for the observer viewing position and the observer viewing orientation from the scene data.
Method and apparatus of encoding/decoding image data based on tree structure-based block division
Disclosed are methods and apparatuses for image data encoding/decoding. A method of decoding an image includes receiving a bitstream in which the image is encoded; obtaining index information for specifying a block division type of a current block in the image; and determining the block division type of the current block from a candidate group pre-defined in the decoding apparatus. The candidate group includes a plurality of candidate division types, including at least one of a non-division, a first quad-division, a second quad-division, a binary-division or a triple-division. The method also includes dividing the current block into a plurality of sub-blocks; and decoding each of the sub-blocks with reference to syntax information obtained from the bitstream.
Automatic composition of composite images or videos from frames captured with moving camera
A processing device generates composite images from a sequence of images. The composite images may be used as frames of video. A foreground/background segmentation is performed at selected frames to extract a plurality of foreground object images depicting a foreground object at different locations as it moves across a scene. The foreground object images are stored to a foreground object list. The foreground object images in the foreground object list are overlaid onto subsequent video frames that follow the respective frames from which they were extracted, thereby generating a composite video.
DETECTION DEVICE AND IMAGE DISPLAY MODULE
A detection device includes a camera and a detector. The camera captures an image of a human face. The detector detects a position of a human eye based on the captured image output from the camera by template matching.
SAMPLING ACROSS MULTIPLE VIEWS IN SUPERSAMPLING OPERATION
Sampling across multiple views in supersampling operation is described. An example of an apparatus includes one or more processing resources configured to perform a supersampling operation for image data generated for multiple views utilizing one or more neural networks, the processing resources including at least a first circuitry to process a first current frame including first image data for a first view, and a second circuitry to process a second current frame including second image data for a second view, the first view and second view being displaced from each other, the processing resources to receive for processing the first current frame and the second current frame, and perform supersampling processing utilizing the one or more neural networks based on at least the first current frame and the second current frame and one or more prior generated frames for each of the views.
Method and apparatus for displaying stereoscopic information related to ultrasound sectional plane of target object
A method of displaying stereoscopic information related to an ultrasound sectional plane of a target object includes setting a line of interest on the ultrasound sectional plane of the target object based on a received input; obtaining an ultrasound signal of the ultrasound sectional plane of the target object along the set line of interest; converting the obtained ultrasound signal to represent the stereoscopic information in a three-dimensional manner; and displaying the stereoscopic information related to the ultrasound sectional plane of the target object.
Sensor signal visualization for sensors of coordinate measuring machines and microscopes
Sensor signals from a sensor of a coordinate measuring machine or microscope describe a workpiece arranged within a space. The sensor and the space are movable relative to one another. A method of visualizing the sensor signals includes obtaining data relating to a three-dimensional scene that is stationary relative to the space. The method includes generating a two-dimensional view image of the scene. The view image has opposing edges predefined with respect to at least one of the two directions. A central region of the view image is located between the edges. The method includes, repeatedly, obtaining a two-dimensional sensor representation of the workpiece and combining the sensor representation with the view image to form a two-dimensional output image. The method includes, in response to movement between the sensor and the space, generating a new view image if the central region would extend beyond either of the edges.