Patent classifications
H04N23/16
CONFIGURABLE PLATFORM
A fluorescence imaging system for imaging an object, the system includes a white light provider that emits white light, an excitation light provider that emits excitation light in a plurality of excitation wavebands for causing the object to emit fluorescent light, a component that directs the white light and excitation light to the object and collects reflected white light and emitted fluorescent light from the object, a filter that blocks light in the excitation wavebands and transmits at least a portion of the reflected white light and fluorescent light, and an image sensor assembly that receives the transmitted reflected white light and the fluorescent light.
SYSTEMS AND METHODS FOR GENERATING A DIGITAL IMAGE
A system, method, and computer program product for generating a digital image is disclosed. In use, a first image and a second image are received from a first image sensor, where the first image sensor detects wavelengths of a visible spectrum. A third image and a fourth image are received from a second image sensor, where the second image sensor detects wavelengths of a non-visible spectrum. Using an image processing subsystem, a resulting image is generated by combining at least three of: the first image, the second image, the third image, or the fourth image.
IMAGE ACQUISITION SYSTEM AND IMAGE ACQUISITION METHOD
An image acquisition system includes: a first narrowband light source that emits first narrowband light for exciting a luminescent agent that exists in an observation target and emits light having a wavelength belonging to a visible light wavelength band; a second narrowband light source that emits second narrowband light in a wavelength band of ±30 nm of a peak light emission wavelength of the luminescent agent; a broadband light source that emits broadband light for illuminating the observation target; a first image sensor on which an image of light in a light emission wavelength band including a wavelength corresponding to light emitted from the luminescent agent is formed; and a second image sensor including one or more image sensors on which an image of light in a wavelength band other than the light emission wavelength band is formed.
High resolution thin multi-aperture imaging systems
A multi-aperture imaging system comprising a first camera with a first sensor that captures a first image and a second camera with a second sensor that captures a second image, the two cameras having either identical or different FOVs. The first sensor may have a standard color filter array (CFA) covering one sensor section and a non-standard color CFA covering another. The second sensor may have either Clear or standard CFA covered sections. Either image may be chosen to be a primary or an auxiliary image, based on a zoom factor. An output image with a point of view determined by the primary image is obtained by registering the auxiliary image to the primary image.
Computer-vision-based object tracking and guidance module
An apparatus comprises a mount body by which the apparatus is secured to a structure. A camera assembly includes an image sensor adapted to capture images within its field of view. A lighting assembly houses one or more light sources including a directional light source. A control-board assembly fixed to the mount body, houses control boards including one or more processors configured to acquire information about an object, to associate a location within the field of view of the image sensor with the object, to point light emitted by the directional light source at the location associated with the obj ect by rotating the lighting assembly and turning the laser assembly, and, based on an image acquired from the camera assembly, to detect change within the field of view of the image sensor corresponding to placement or removal of the object.
3 MOS camera
A 3 MOS camera includes a first prism that has a first reflection film which reflects IR light that causes a first image sensor to receive the IR light, a second prism that has a second reflection film which reflects A % (A: a predetermined real number) visible light and that causes a second image sensor to receive the A % visible light, a third prism that causes a third image sensor to receive a (100−A) % visible light, and a video signal processor that combines a first video signal, a second video signal, and a third video signal of an observation part. The video signal processor performs pixel shifting on one of the second video signal and the third video signal having substantially same brightness to generate a fourth video signal and outputs a video signal obtained by combining the fourth video signal and the first video signal.
SYSTEM AND APPARATUS OF UNDER-DISPLAY CAMERA
A device including a display including a main portion, a first region and a second region being in different positions within a boundary of the main portion, the first region having a first structure of light-blocking, and the second region having a second structure of light-blocking elements, a first light sensor positioned under the first region, the first light sensor being configured to capture a first image having at least one first characteristic, wherein the first characteristic depends on the first structure, a second light sensor positioned under the second region, the second light sensor being configured to capture a second image having at least one second characteristic, wherein the second characteristic depends on the second structure, and a memory including code that when executed by a processor causes the processor to generate a third image based on the first image and the second image.
SYSTEM AND APPARATUS OF UNDER-DISPLAY CAMERA
A device including a display including a main portion, a first region and a second region being in different positions within a boundary of the main portion, the first region having a first structure of light-blocking, and the second region having a second structure of light-blocking elements, a first light sensor positioned under the first region, the first light sensor being configured to capture a first image having at least one first characteristic, wherein the first characteristic depends on the first structure, a second light sensor positioned under the second region, the second light sensor being configured to capture a second image having at least one second characteristic, wherein the second characteristic depends on the second structure, and a memory including code that when executed by a processor causes the processor to generate a third image based on the first image and the second image.
Using IR Sensor With Beam Splitter to Obtain Depth
Obtaining depth information using an IR sensor with a beam splitter including illuminating a subject with IR light using an IR light source; receiving reflected light including visible light and the IR light at a beam splitter; splitting the reflected light into two identical beams, a first beam and a second beam, using the beam splitter; receiving and processing the first beam at an IR sensor to pass the IR light and to block the visible light, to generate an IR image; receiving and processing the second beam at a visible light sensor to pass the visible light and to block the IR light, to generate a visible light image; and using a time of flight of the IR light transmitted by the IR light source and received by the IR sensor to calculate a distance of the subject from the beam splitter.
EXTRACTION OF SUBJECT FROM BACKGROUND
Video capture of a subject, including: capturing video data of the subject using a first camera; capturing infrared (IR) video data of the subject using a second camera linked to the first camera, in such a manner that the first and second camera share the same field of view, wherein the second camera is sensitive to IR light for capturing the IP video data of the subject; illuminating the subject with at least one IR light source; and processing the video data from the first camera and the IR video data from the second camera to identify an outline of the illuminated subject in the video data using the IR video data.