Patent classifications
G06V10/145
REARVIEW ASSEMBLY INCORPORATING HIGH-TRANSMITTANCE NIR FILM
A rearview assembly includes an electrochromic element. The electrochromic element includes a first substrate including a first surface and a second surface. The electrochromic element further includes a second substrate comprising a third surface and a fourth surface. The first substrate and the second substrate form a cavity. The electrochromic element includes an electrochromic medium contained in the cavity. The rearview assembly includes an image sensor directed toward the fourth surface and configured to capture near-infrared light reflected from an object and projected through the electrochromic element. The rearview assembly includes a transflective film disposed adjacent to the fourth surface having a near-infrared light transmission level and a visible light reflectance level.
Visual, depth and micro-vibration data extraction using a unified imaging device
A unified imaging device used for detecting and classifying objects in a scene including motion and micro-vibrations by receiving a plurality of images of the scene captured by an imaging sensor of the unified imaging device comprising a light source adapted to project on the scene a predefined structured light pattern constructed of a plurality of diffused light elements, classifying object(s) present in the scene by visually analyzing the image(s), extracting depth data of the object(s) by analyzing position of diffused light element(s) reflected from the object(s), identifying micro-vibration(s) of the object(s) by analyzing a change in a speckle pattern of the reflected diffused light element(s) in at least some consecutive images and outputting the classification, the depth data and data of the one or more micro-vibrations which are derived from the analyses of images captured by the imaging sensor and are hence inherently registered in a common coordinate system.
Visual, depth and micro-vibration data extraction using a unified imaging device
A unified imaging device used for detecting and classifying objects in a scene including motion and micro-vibrations by receiving a plurality of images of the scene captured by an imaging sensor of the unified imaging device comprising a light source adapted to project on the scene a predefined structured light pattern constructed of a plurality of diffused light elements, classifying object(s) present in the scene by visually analyzing the image(s), extracting depth data of the object(s) by analyzing position of diffused light element(s) reflected from the object(s), identifying micro-vibration(s) of the object(s) by analyzing a change in a speckle pattern of the reflected diffused light element(s) in at least some consecutive images and outputting the classification, the depth data and data of the one or more micro-vibrations which are derived from the analyses of images captured by the imaging sensor and are hence inherently registered in a common coordinate system.
Multi-channel depth estimation using census transforms
A depth estimation system is described capable of determining depth information using two images from two cameras. A first camera captures a first image and a second camera captures a second image, both images including a plurality of light channels. A scan direction is selected from a plurality of scan directions. For the selected scan direction, along each of a plurality of scanlines, the system compares pixels from the first image to pixels from the second image. The comparison is based on calculating a census transform for each pixel in the first image and a census transform for each pixel in the second image. This comparison is used to determine a stereo correspondence between the pixels in the first image and the pixels in the second image. The system generates a depth map based on the stereo correspondence.
Multi-channel depth estimation using census transforms
A depth estimation system is described capable of determining depth information using two images from two cameras. A first camera captures a first image and a second camera captures a second image, both images including a plurality of light channels. A scan direction is selected from a plurality of scan directions. For the selected scan direction, along each of a plurality of scanlines, the system compares pixels from the first image to pixels from the second image. The comparison is based on calculating a census transform for each pixel in the first image and a census transform for each pixel in the second image. This comparison is used to determine a stereo correspondence between the pixels in the first image and the pixels in the second image. The system generates a depth map based on the stereo correspondence.
Light emitting device
A light emitting device includes a wiring substrate, a light emitting element array that includes a first side surface and a second side surface facing each other, and a third side surface and a fourth side surface connecting the first side surface and the second side surface to each other and facing each other, the light emitting element array being provided on the wiring substrate, a driving element that is provided on the wiring substrate on the first side surface side and drives the light emitting element array, a first circuit element and a second circuit element that are provided on the wiring substrate on the second side surface side to be arranged in a direction along the second side surface, and a wiring member that is provided on the third side surface side and the fourth side surface side and extends from a top electrode of the light emitting element array toward an outside of the light emitting element array.
Light emitting device
A light emitting device includes a wiring substrate, a light emitting element array that includes a first side surface and a second side surface facing each other, and a third side surface and a fourth side surface connecting the first side surface and the second side surface to each other and facing each other, the light emitting element array being provided on the wiring substrate, a driving element that is provided on the wiring substrate on the first side surface side and drives the light emitting element array, a first circuit element and a second circuit element that are provided on the wiring substrate on the second side surface side to be arranged in a direction along the second side surface, and a wiring member that is provided on the third side surface side and the fourth side surface side and extends from a top electrode of the light emitting element array toward an outside of the light emitting element array.
LIGHT-EMITTING USER INPUT DEVICE FOR CALIBRATION OR PAIRING
A light emitting user input device can include a touch sensitive portion configured to accept user input (e.g., from a user's thumb) and a light emitting portion configured to output a light pattern. The light pattern can be used to assist the user in interacting with the user input device. Examples include emulating a multi-degree-of-freedom controller, indicating scrolling or swiping actions, indicating presence of objects nearby the device, indicating receipt of notifications, assisting pairing the user input device with another device, or assisting calibrating the user input device. The light emitting user input device can be used to provide user input to a wearable device, such as, e.g., a head mounted display device.
LIGHT-EMITTING USER INPUT DEVICE FOR CALIBRATION OR PAIRING
A light emitting user input device can include a touch sensitive portion configured to accept user input (e.g., from a user's thumb) and a light emitting portion configured to output a light pattern. The light pattern can be used to assist the user in interacting with the user input device. Examples include emulating a multi-degree-of-freedom controller, indicating scrolling or swiping actions, indicating presence of objects nearby the device, indicating receipt of notifications, assisting pairing the user input device with another device, or assisting calibrating the user input device. The light emitting user input device can be used to provide user input to a wearable device, such as, e.g., a head mounted display device.
DUAL-PATTERN OPTICAL 3D DIMENSIONING
An optical dimensioning system includes one or more light emitting assemblies configured to project one or more predetermined patterns on an object; an imaging assembly configured to sense light scattered and/or reflected off the object, and to capture an image of the object while the patterns are projected; and a processing assembly configured to analyze the image of the object to determine one or more dimension parameters of the object. The light emitting assembly may include a single piece optical component configured for producing a first pattern and second pattern. The patterns may be distinguishable based on directional filtering, feature detection, feature shift detection, or the like. A method for optical dimensioning includes illuminating an object with at least two detectable patterns; and calculating dimensions of the object by analyzing pattern separate of the elements comprising the projected patterns. One or more pattern generators may produce the patterns.