H04N2013/0074

Book scanning using machine-trained model
11030488 · 2021-06-08 · ·

This application discloses a technology for flattening a photographed page of a book and straightening texts therein. The technology uses one or more mathematical models to represent a curved shape of the photographed page with certain parameters. The technology also uses one or more photographic image processing techniques to dewarp the photographed page using the parameters of the curved shape. The technology uses one or more additional parameters that represent certain features of the photographed page to dewarp the photographed page.

Method of determining the boundary of a driveable space
10972711 · 2021-04-06 · ·

A method of determining the characteristics of a scene around a vehicle comprises capturing a stereo pair of images of the scene and processing the images to produce a depth map of the scene. Each pixel in the depth map is assigned a value that corresponds to the range of a corresponding region in the scene, the pixels being arranged in a grid of rows and columns with each column of pixels in the grid corresponding to a vertically oriented set of regions in the scene and each row a horizontally oriented set of regions in the scene. The values for one or more columns of pixels in the depth map are binned to form a corresponding histogram, each bin in the histogram having a count value that corresponds to the number of pixels in the column that have a depth within the range assigned to the bin. The or each of the range bin histograms are processed to determine for the or each histogram the lowest range bin that is indicative that an object that represents a non-drivable region is present at a depth that lies in the range of depths assigned to the bin, thereby identifying the location of one or more boundary points that lie on a boundary of a drivable space in the scene.

THREE-DIMENSIONAL SCANNER HAVING SENSORS WITH OVERLAPPING FIELDS OF VIEW
20230408253 · 2023-12-21 ·

A system includes a projector configured to project a plurality of non-coded elements onto an object, the projector having a first optical axis. The system includes a first camera having a first lens and a first sensor. The first lens defines a second optical axis. The system includes a second camera having a second lens and a second sensor. The second lens defines a third optical axis. The projector, the first camera, and the second camera are disposed on a substantially straight line in a first direction. The first optical axis is substantially parallel to the second optical axis, which is substantially parallel to the third optical axis. A center of the first sensor is displaced along the first direction away from the second optical axis, and a center of the second sensor is displaced along the first direction away from the third optical axis.

Methods and systems for content processing

Mobile phones and other portable devices are equipped with a variety of technologies by which existing functionality can be improved, and new functionality can be provided. Some aspects relate to visual search capabilities, and determining appropriate actions responsive to different image inputs. Others relate to processing of image data. Still others concern metadata generation, processing, and representation. Yet others concern user interface improvements. Other aspects relate to imaging architectures, in which a mobile phone's image sensor is one in a chain of stages that successively act on packetized instructions/data, to capture and later process imagery. Still other aspects relate to distribution of processing tasks between the mobile device and remote resources (the cloud). Elemental image processing (e.g., simple filtering and edge detection) can be performed on the mobile phone, while other operations can be referred out to remote service providers. The remote service providers can be selected using techniques such as reverse auctions, through which they compete for processing tasks. A great number of other features and arrangements are also detailed.

System and method for simultaneous consideration of edges and normals in image features by a vision system

This invention applies dynamic weighting between a point-to-plane and point-to-edge metric on a per-edge basis in an acquired image using a vision system. This allows an applied ICP technique to be significantly more robust to a variety of object geometries and/or occlusions. A system and method herein provides an energy function that is minimized to generate candidate 3D poses for use in alignment of runtime 3D image data of an object with model 3D image data. Since normals are much more accurate than edges, the use of normal is desirable when possible. However, in some use cases, such as a plane, edges provide information in relative directions the normals do not. Hence the system and method defines a normal information matrix, which represents the directions in which sufficient information is present. Performing (e.g.) a principal component analysis (PCA) on this matrix provides a basis for the available information.

Apparatus, method and computer program for performing object recognition

An apparatus for performing object recognition includes an image camera to capture a first resolution image and a depth map camera to capture a second resolution depth map. The first resolution is greater than the second resolution. The apparatus is configured to perform object recognition based on the image and the depth map.

Image processing method and apparatus, and electronic device
10908492 · 2021-02-02 · ·

This application provides an image processing method and an electronic device. The method includes: obtaining a first image and a second image that are acquired at a same acquisition moment, where the first image is obtained by using a first photographing unit to acquire an image in a first acquisition region, and the second image is obtained by using a second photographing unit to acquire an image in a second acquisition region, where a parameter of the first photographing unit and a parameter of the second photographing unit are the same, and there is an overlapping acquisition region between the first acquisition region and the second acquisition region; and determining, according to the first image and the second image, whether the first photographing unit and the second photographing unit are normal.

Generating composite stereoscopic images usually visually-demarked regions of surfaces
11057612 · 2021-07-06 · ·

A system, method or compute program product for generating composite images. One of the systems includes a capture device to capture an image of a physical environment; and one or more storage devices storing instructions that are operable, when executed by one or more processors of the system, to cause the one or more processors to: obtain an image of the physical environment as captured by the capture device, identify a visually-demarked region on a surface in the physical environment as depicted in the image, process the image to generate a composite image of the physical environment that includes a depiction of a virtual object, wherein a location of the depiction of the virtual object in the composite image is based on a location of the depiction of the visually-demarked region in the image, and cause the composite image to be displayed for a user.

3D SCENE MODELLING SYSTEM BY MULTI-VIEW PHOTOGRAMMETRY
20200408519 · 2020-12-31 ·

A 3D modelling system for three-dimensionally modelling a scene by multi-view photogrammetry has cameras placed around the scene and grouped together in pairs that are spaced further apart from one another than the two cameras in each pair are spaced apart from each other, and a digital processing device that is configured to produce modelling data by applying stereoscopy processing by comparison firstly between the images of the scene that are produced by respective ones of the two cameras of a first and a second pair of cameras and secondly between the images produced by two cameras belonging to respective ones of the first and second pairs of cameras, the first and the second pairs of cameras being arranged respectively in a horizontal plane and in a vertical plane.

Electronic device and method for tiled video multi-channel playback

An electronic device and a method for a tiled video multi-channel playback are disclosed. The electronic device includes a device communication unit and a device controller. The device communication unit establishes a communication channel with a streaming server. The device controller can detect a region of interest (ROI) of an image related to a media presentation description (MPD) tile provided by the streaming server, transmit ROI information to the streaming server and receive tiles of a tiled video having a first resolution and corresponding to the ROI information, a full video having a second resolution lower than the first resolution, and a plurality of 2D videos from the streaming server. The device controller can also generate a 360-degree VR video by synthesizing the video tiles and the full video, and control a simultaneous playback of the generated 360-degree VR video and the plurality of 2D videos.