Ultra-wide field-of-view flat optics
10979635 ยท 2021-04-13
Assignee
Inventors
Cpc classification
G02B13/06
PHYSICS
G02B1/002
PHYSICS
G02B6/0023
PHYSICS
International classification
G02B1/00
PHYSICS
G02B13/06
PHYSICS
Abstract
Wide-angle optical functionality is beneficial for imaging and image projection devices. Conventionally, wide-angle operation is attained by a complicated assembly of optical elements. Recent advances have led to meta-surface lenses or meta-lenses, which are ultra-thin planar lenses with nanoantennas that control the phase, amplitude, and/or polarization of light. Here, we present a meta-lens capable of diffraction-limited focusing and imaging over an unprecedented >170 angular field of view (FOV). The lens is integrated on a one-piece flat substrate and includes an aperture on one side and a single meta-surface on the other side. The meta-surface corrects third-order Seidel aberrations, including coma, astigmatism, and field curvature. The meta-lens has a planar focal plane, which enables considerably simplified system architectures for imaging and projection. The meta-lens design is generic and can be readily adapted to different meta-atom geometries and wavelength ranges to meet diverse application demands.
Claims
1. An optical component comprising: a substrate; an opaque layer on a first side of the substrate, the opaque layer defining an aperture to transmit light over a field of view of at least 120; and a meta-surface, on a second side of the substrate opposite the first side, to focus the light transmitted by the aperture through the substrate, wherein the meta-surface comprises an array of at least 10001000 meta-atoms.
2. The optical component of claim 1, wherein the substrate is planar.
3. The optical component of claim 1, wherein the substrate is curved.
4. The optical component of claim 1, wherein the substrate has a thickness t.sub.sub and a refractive index at a wavelength of the light of n.sub.sub, the aperture has a diameter of D.sub.in, and the meta-surface has a diameter D.sub.meta=D.sub.in+t.sub.sub tan[sin.sup.1(1/n.sub.sub)].
5. The optical component of claim 1, wherein the meta-surface is configured to focus the light incident over all of the field of view to a focal plane parallel to the second side of the substrate.
6. The optical component of claim 5, further comprising: a detector array, in the focal plane, to detect the light focused by the meta-surface.
7. The optical component of claim 5, further comprising: a light source array in the focal plane, and wherein the meta-surface is configured to collimate a beam emitted by the light source array and the aperture is configured to emit the beam.
8. The optical component of claim 7, wherein the meta-surface and/or the aperture are further configured to modulate the beam emitted by the light source array.
9. The optical component of claim 1, wherein the meta-surface is configured to focus the light incident over all of the field of view with a Strehl ratio of at least 80%.
10. The optical component of claim 1, wherein the meta-surface is a first meta-surface, and further comprising: a second meta-surface, disposed in at least a portion of the aperture, to modulate and/or filter the light transmitted by the aperture.
11. The optical component of claim 10, wherein the second meta-surface is configured to modulate the light transmitted by the aperture with a spatial modulation pattern that depends on an angle of incidence of the light transmitted by the aperture.
12. An optical component comprising: a substrate; an opaque layer on a first side of the substrate, the opaque layer defining an aperture to transmit light over a field of view of at least 120; and a meta-surface, on a second side of the substrate opposite the first side, to focus the light transmitted by the aperture through the substrate, wherein the meta-surface has different yet continuous portions configured to capture input beams at different angles of incidence.
13. An optical component comprising: a substrate; an opaque layer on a first side of the substrate, the opaque layer defining an aperture to transmit light over a field of view of at least 120; and a meta-surface, on a second side of the substrate opposite the first side, to focus the light transmitted by the aperture through the substrate, wherein the meta-surface is configured to correct at least one third-order Seidel aberration.
14. A method comprising: transmitting light over a field of view of at least 120 through an aperture formed on a first side of a substrate; and focusing the light with a meta-surface on a second side of the substrate opposite the first side, wherein focusing the light comprises correcting at least one third-order Seidel aberration.
15. The method of claim 14, wherein focusing the light comprises focusing the light incident over all of the field of view to a focal plane parallel to the second side of the substrate.
16. The method of claim 15, further comprising: detecting the light focused by the meta-surface with a detector array in the focal plane.
17. The method of claim 15, further comprising: collimating a beam emitted by the light source array in the focal plane with the meta-surface; and emitting the beam through the aperture.
18. The method of claim 14, wherein focusing the light comprises focusing the light incident over all of the field of view with a Strehl ratio of at least 80%.
19. The method of claim 14, further comprising: modulating at least of a phase, an amplitude, a polarization, or a wavelength of the light transmitted by the aperture with another meta-surface disposed in at least a portion of the aperture based on an angle of incidence of the light transmitted by the aperture.
20. The method of claim 14, further comprising: filtering the light transmitted by the aperture with another meta-surface disposed in at least a portion of the aperture.
21. A sensor comprising: a substrate; a light source array, supported by a first portion of the substrate, to emit light; a first meta-lens, in optical communication with the light source array, to project the light emitted by the light source array toward an object over a field of view of at least about 120, the first meta-lens comprising a first planar substrate having a first meta-surface on a first side facing the light source array and a first aperture on a second side; a second meta-lens, in optical communication with the object, to collect the light scattered and/or reflected by the object over a field of view of at least about 120, the second meta-lens comprising a second planar substrate having a second aperture on a first side facing the object and a second meta-surface on a second side; and a detector array, supported by a second portion of the substrate and in optical communication with the second meta-lens, to detect the light collected by the second meta-lens.
Description
BRIEF DESCRIPTIONS OF THE DRAWINGS
(1) The skilled artisan will understand that the drawings primarily are for illustrative purposes and are not intended to limit the scope of the inventive subject matter described herein. The drawings are not necessarily to scale; in some instances, various aspects of the inventive subject matter disclosed herein may be shown exaggerated or enlarged in the drawings to facilitate an understanding of different features. In the drawings, like reference characters generally refer to like features (e.g., functionally and/or structurally similar elements).
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
(15)
(16)
(17)
(18)
(19)
(20)
(21)
(22)
(23)
(24)
(25)
(26)
(27)
(28)
(29)
(30)
(31)
DETAILED DESCRIPTION
(32) An inventive wide field-of-view (WFOV) meta-lens has a simple and easy-to-fabricate configuration, with a single meta-surface layer and an aperture integrated on opposite sides of a single thin substrate. It can have a diffraction-limited FOV exceeding 170 and a planar focal plane, which significantly simplifies the associated detector (for imaging and detection) or light emitter (for image/beam projection, display, etc.) array design. It can operate at a wide range of wavelengths (e.g., from the visible to the infrared (IR)), depending on the design of the meta-surface and the substrate and meta-surface materials. A meta-surface can be designed to operate at any wavelength from the microwave to ultraviolet (UV) regions of the electromagnetic spectrum, with a bandwidth that spans up to an octave.
(33) Moreover, the WFOV meta-lens design principles disclosed here are generic and applicable to arbitrary meta-atom configurations and wavelength ranges to meet diverse application demands. With a proper meta-surface, a meta-lens can be a panoramic meta-lens with broadband operation and polarization diversity. As another example, nano-rod or nano-pillars can be used for polarization-insensitive designs. A variety of meta-surface materials systems can be chosen for desired wavelength ranges, including, for example, PbTe and CaF.sub.2, and a-Si and Al.sub.2O.sub.3.
(34) A WFOV meta-lens can be implemented using a Huygens meta-surface with an ultra-thin meta-atom profile. A Huygens meta-surface can be fabricated easily but may constrain sensitivity to wavelength and polarization. In any event, the WFOV design described herein is generic and applicable to arbitrary meta-atom configurations. With a proper choice of meta-atoms, the meta-lens can perform panoramic imaging (e.g., a FOV exceeding 170) over a broadband with polarization diversity.
(35) Thanks to their extremely WFOVs, planar surfaces, and flat focal plane, inventive meta-lenses are particularly well suited for sensing, image projection, lidar, imaging, optical projection, augmented reality/virtual reality, beam steering, and 3-D sensing applications. For example, the beam projection function of an inventive device can be used for LIDAR systems, when the light emitters are individually modulated or switched on and off to steer, switch, or tune the output beams for wide-angle illumination. The same WFOV meta-lens can be used in the detection module for wide-angle signal collection.
(36) A Wide Field-of-View (WFOV) Meta-Lens
(37)
(38) The substrate 110 may be made of any suitable material that transmits light at the meta-lens's operating wavelength. The substrate 110 may be rigid, flexible, or stretchable. It can be flat/planar on both sides as shown in
(39) The meta-surface 120 includes an array of sub-wavelength optical structures (also known as meta-atoms; described and shown below) that modify the amplitude, phase, and/or polarization of incoming wave fronts. These meta-atoms can have the same or different shapes, sizes, and orientations. For example, they can be rectangular, cylindrical, H-shaped, or L-shaped. They are arrayed on a lattice with a pitch that is less than or equal to the operating wavelength of the meta-lens 100. The lattice can have any suitable period and shape (e.g., square, rectangular, or hexagonal). The lattice can also be aperiodic, for example, with spacing defined by the gap between adjacent meta-atoms, such as a constant gap distance, or with randomly positioned meta-atoms, which are patterned to provide a desired phase profile over the entire meta-surface 120.
(40) The meta-atoms are made of transparent material (e.g., the same material as the substrate 110, with the meta-atoms formed by patterning one side of the substrate 110). The meta-atoms' shapes, sizes, and layout can be selected so that the meta-surface's spectral response does not change with AOIthe optical phase/amplitude change imparted by the meta-atoms with varying AOI is automatically accounted for in their design. The meta-surface 120 can also be designed for rotationally asymmetric focusing (e.g., focusing at some AOIs but not others).
(41) (Alternatively, the meta-surface 120 can be replaced by a meta-material, multi-layer meta-surfaces, or diffractive optical element (DOE) that provides the same or similar effective phase profile. For instance, a DOE could be implemented as a binary or multi-level grayscale DOE with a feature size that is larger than the lens's operating wavelength. Similarly, the overall lens architecture works for diffractive optical lenses.)
(42) The aperture stop 130 is defined by a layer 132 of opaque (e.g., absorptive or reflective) material on the upper surface 112 of the substrate 110. The upper surface 112 can also define or be partially or fully covered with a meta-surface (not shown) that modulates the intensity and/or phase of the incident light to form an effective aperture stop. (Alternatively, this meta-surface can be replaced by a metamaterial, multi-layer meta-surfaces, or DOE.) The aperture stop 130 can be circular with a diameter given by:
D.sub.in=D.sub.meta2t.sub.sub tan[sin.sup.1(1/n.sub.sub)]
This diameter can range from microns to millimeters, with a numerical aperture (NA) that ranges from 0 to 1. The numerical aperture can higher (e.g., 1.5) if the meta-lens is immersed in oil or other high-index material.
(43) The aperture can also be square, elliptical, hexagonal, rectangular, or any other suitable shape. Alternatively, the aperture can include one or more sub-regions, patches, or arrays configured to modulate or encode the input light in spectrum, phase, amplitude, polarization, etc. For example, at least a portion of the aperture 130 may be patterned with another meta-surface 134 that filters light passing through the aperture 130. If desired, the edge of the aperture stop 130 can be apodized, e.g., with a Gaussian or super-Gaussian apodization, to reduce deleterious edge effects.
(44) By spatially decoupling the meta-surface 120 and aperture stop 130 while positioning them on the same substrate, the meta-lens 100 can capture input beams at different angles of incidence (AOIs) on different yet continuous portions of the meta-surface 120, facilitating local optimization of the phase profiles, e.g., by optimizing against a figure of merit that accounts for focusing quality at multiple AOIs. The meta-surface phase profile is designed so that the root-mean-square (RMS) wave front error from an ideal spherical wave front over the input aperture is always smaller than 0.0745 wavelengths. This ensures that the meta-lens 100 has a Strehl ratio of over 80% over its entire field-of-view, which can be 120, 130, 140, 150, 160, 170, 175, 179, or nearly 180 for a flat substrate, thereby achieving diffraction-limited performance at various light illumination conditions. For meta-lens with a curved, bent, or warped substrate, the field-of-view can be even larger than 180. Similar to the flat surface case, the incident light is refracted (or diffracted using a meta-surface) from the input aperture to the backside meta-surface. Conventional fisheye lenses achieve FOV>180 using a largely curved front lens.
(45) The meta-lens 100 can operate at any of a variety of wavelengths, depending on its dimensions, the meta-surface design, and the substrate material. For example, a meta-lens 100 designed to operate at a wavelength of 5.2 m may have a 2-mm-thick calcium fluoride (CaF.sub.2) planar substrate 110 (n.sub.sub=1.4 at 5.2 m) with a 1-mm-diameter circular aperture 130 and a 5.2 mm5.2 mm meta-surface 120. This meta-surface 120 may contain an array of 2,0002,000 Huygens meta-atoms made of PbTe with a square lattice constant of 2.5 m. The meta-surface 120 can have a constant focal length of 2 mm, corresponding to an effective numerical aperture (NA) of 0.24. At an incident angle of nearly 90, the maximum angle of light propagation inside the substrate is 45.7. As shown below, the phase response of the meta-atoms making up the meta-surface 120 depends only weakly on the beam incident angles within meta-lens's WFOV.
(46) The meta-lens 100 operates differently than a meta-lens designed with separated angular channels, in which non-overlapping regions on a meta-surface are dedicated to beams at different AOIs. Because it has dedicated non-overlapping regions, a meta-lens with separated angular channels can only achieve high-quality focusing for a discrete set of incident angles. In an inventive meta-lens 100, the judiciously designed meta-surface phase profile and meta-lens architecture allow diffraction-limited focusing of beams with continuously varying incident angles and mutually overlapping beam profiles on the meta-surface side 114 of the substrate 110. Therefore, the meta-lens 100 can achieve aberration-free beam focusing or beam collimation and thus image projection for any light direction from or to any point on the front hemisphere.
(47) In addition to correcting aberrations such as coma and astigmatism, the meta-lens 100 features a planar focal plane 141 across the entire FOV. The elimination of Petzval field curvature is beneficial in a wide range of applications, including imaging and image projection, by facilitating standard planar detector or emitter array integration. For example,
(48) Meta-Lens Meta-Surface Design and Modeling
(49) A meta-lens can be designed utilizing a hierarchical combination of full-wave simulations (such as finite-element method (FEM), finite-difference time-domain (FDTD) method, and finite integration technique (FIT)) and Kirchhoff diffraction integral. At the sub-wavelength-scale, full-wave simulations can be used to design and model the meta-atoms in the meta-surface for desired optical responses. At the macroscopic system level, the diffraction integral method incorporating the full wave simulation results enables computationally efficient validation of the focusing characteristics of the entire meta-lens and can be used to optimize the phase profile of the meta-surface.
(50)
(51)
(52) TABLE-US-00001 TABLE 1 Dimensions of meta-atoms used in the meta-optical devices. Meta-atom number 1 2 3 4 5 6 7 8 L.sub.x (L.sub.xd) 2.0 1.8 2.0 2 1.78 1.38 0.62 2 (0.6) (0.6) (0.7) L.sub.y (L.sub.yd) 1.7 1.9 1.6 0.78 0.7 0.66 0.52 1.26 (0.9) (0.8) (0.7)
(53)
(54)
(55)
(56) The meta-atom simulations illustrated in
(57) The focusing and imaging behavior of the WFOV meta-lens was modeled following the Kirchhoff diffraction integral, a physically rigorous form of the Huygens-Fresnel principle. The model starts with computing the Huygens point spread function of the optical system. It incorporates angular-dependent phase profiles at the meta-surface and propagates wave fronts emitted from each meta-atoms with corresponding amplitude and phase to the image plane where its complex amplitude is derived. The diffraction of the wave front through space is given by the interference or coherent sum of the wave fronts from the Huygens sources. The intensity at each point on the image plane is the square of the resulting complex amplitude sum.
(58) The initial optical structure and phase profile of the meta-lens can be designed using OpticStudio (Zemax, LLC). An analytical model based on the Kirchhoff diffraction integral can be subsequently utilized to analyze the full meta-surface performance under different AOIs. The analytical model incorporates angular-dependent phase masks following individual meta-atom responses under different AOIs obtained from full-wave simulations (e.g., as in
(59)
where w.sub.AOI(i) gives the weighting factor at different AOIs. This form of FOM simultaneously examines multiple beams over the entire FOV. For the exemplary design, an initial angular interval of AOI=5 between neighboring beams is sufficient to sample the entire FOV continuously. The optimized meta-lens phase profile is shown in
(60) The commercial optical design software Zemax OpticStudio can be used to obtain the initial phase profile of the meta-surface under ideal conditions. The rotationally symmetric phase profiles are expressed in a polynomial form:
(61)
where () is the desired phase response of specific meta-atoms with ={square root over (x.sup.2+y.sup.2)}, a.sub.n are aspheric coefficients, and R is the normalization radius. TABLES 1 and 2 (below) show the normalization radii and aspheric coefficients for example mid-IR and NIR WFOV meta-lenses, respectively:
(62) TABLE-US-00002 TABLE 1 Phase profile design coefficients for mid-IR WFOV meta-lens R (mm) a1 a2 a3 a4 a5 a6 a7 2.219e3 1492.12 232.31 916.34 2179.49 2829.43 1837.91 467.25
(63) TABLE-US-00003 TABLE 2 Phase profile design coefficients for NIR WFOV meta-lens R (mm) a1 a2 a3 a4 a5 a6 a7 a8 a9 a10 2.25e3 6708.07 63.55 175.6 290.48 264.77 106.33 16.11 34.28 12.98 1.69
(64) Optimization starts with a standard hyperbolic phase profile consistent with the optical system specifications (e.g., focal length and f/#) over a small AOI range as the initial input. Once the initial optimization cycle converges for the starting AOI range, the result is used as the input for the next optimization iteration cycle with an expanded AOI range. The process continues till the final result converges over a targeted AOI range (e.g., 90).
(65) More specifically, numerical optimization using the Levenberg-Marquardt algorithm (also known as the damped least-squares method) is implemented to maximize the merit function (e.g., the FOM defined above) in each optimization cycle. The Kirchhoff diffraction integral is used to numerically calculate the Strehl ratio at each field angle as well as the FOM. Angular-dependent responses of each meta-atoms are also incorporated to generate angular-dependent phase masks after spatial and phase discretization/mapping of the initial phase profile. In each optimization cycle, a FOM with equal weights for all AOIs over the range is initially used and maximized. Maximizing this equal-weight FOM can sometimes result in Strehl ratios less than 0.8 at some field angles. In this case, the optimization is repeated with adjusted weighting factors until the Strehl ratios are above 0.8 for all AOIs within the target range. The final phase profile, shown in
(66)
(67)
(68)
(69)
(70) WFOV Meta-Lens Fabrication
(71)
(72) More specifically, the meta-lens in
(73) One side of the sample was covered with a double-layer photoresist composed of PMGI (800 nm thick) and ZEP 520A (400 nm thick). The PMGI layer was spin-coated at 2400 revolutions per minute (rpm) for 1 minute, then baked at 190 C. for 3 minutes. The baking step is critical for assuring mechanical stability of the PMGI layer. The ZEP layer was spin-coated at 4000 rpm for 1 minute and baked at 190 C. for 2 minutes. To prevent charging effects while performing electron beam (e-beam) lithography, the sample was covered with a water-soluble conductive polymer (ESpacer 300Z, Showa Denko America, Inc.) and placed a conducting clamp on top of the substrate.
(74) The meta-surface patterns (
(75) After photoresist development, a 650-nm-thick PbTe film was deposited by thermal evaporation (custom-designed system, PVD Products, Inc.) at a rate of 17 /s and a base pressure of 10.sup.6 Torr. Before deposition, the sample was pre-cleaned with oxygen plasma to improve adhesion of the film. Later, the meta-surface pattern was transferred by lifting off the material on top of the photoresist by overnight soaking in N-Methyl-2-pyrrolidone (NMP).
(76) The other side of the sample was patterned with a circular aperture of 1 mm in diameter. The side patterned with the PbTe meta-surface was protected by a dry film photoresist (DuPont MX5000 series) during the aperture fabrication. To fabricate the aperture, the surface was cleaned with oxygen plasma and spin-coated with a negative photoresist NR1000PY (Futurrex, Inc.) at 1500 rpm for 1 minute. Then the sample was soft baked at 115 C., exposed to UV light through the mask for 40 seconds, and hard-baked at 155 C. The exposed photoresist was subsequently developed in RD6 for 10 seconds and rinsed with water afterwards. Then a 200 nm layer of tin was deposited by thermal evaporation and lifted off by removing the photoresist with acetone. Finally, the dry film photoresist covering the meta-surface side was removed by overnight NMP treatment.
(77) WFOV Meta-Lens Characterization
(78)
(79)
(80)
(81) The power focused by a meta-lens P.sub.ms,foc(.sub.i) can be expressed in terms of total incident power P.sub.0 transmitted through the frontside aperture (e.g., a 1 mm circular aperture), meta-lens focusing efficiency f(.sub.i), and Fresnel transmittance factor T.sub.p(.sub.i) accounting for reflection losses at the interface between air and the substrate (e.g., CaF.sub.2):
P.sub.ms,foc(.sub.i)=P.sub.0.Math.T.sub.p(.sub.i)f(.sub.i).(3)
The total incident power P.sub.0 can be further written as P.sub.0=P.sub.0(0).Math.cos(.sub.i), where P.sub.0(0) is the total incident power through the aperture at normal incidence (.sub.i=0). The cosine factor comes in because when the same collimated laser beam (with a beam diameter much larger than the aperture size) is incident obliquely on the meta-lens, the power density drops by a factor of cos(.sub.i) due to geometric projection.
(82) In
(.sub.i)=P.sub.ms,foc(.sub.i)/P.sub.ms,trans(.sub.i)(4)
P.sub.ref(.sub.i)=P.sub.0T.sub.P.sup.2(.sub.i)=P.sub.0(0)cos(.sub.i)T.sub.P.sup.2(.sub.i)(5)
(83) In Eq. (5), the T.sub.p(.sub.i) factor is squared because there are two CaF.sub.2-air interfaces with identical transmittance. Finally, the value of focusing efficiency f(.sub.i) is given by:
(84)
(85) P.sub.0, P.sub.ms,trans(.sub.i), and P.sub.ref(.sub.i) were measured using a large-area detector to capture all of the transmitted power. T.sub.p(.sub.i) was then calculated from P.sub.0 and P.sub.ref(.sub.i) according to Eq. (5). P.sub.ms,foc(.sub.i), was quantified by measuring the transmitted power P.sub.hole(.sub.i) incident upon a detector integrated with a 200 m diameter pin hole. The FPA camera imaged the focal plane around the focal spot over a 200 m diameter area. Integrating the optical intensity values from the FPA camera pixel-by-pixel yielded the fraction of power concentrated at the focal spot over the total power transmitted through the pin hole, i.e., P.sub.ms,foc(.sub.i)/P.sub.hole(.sub.i). P.sub.ms,foc(.sub.i) was extracted via P.sub.hole(.sub.i)P.sub.ms,foc(.sub.i)/P.sub.hole(.sub.i). (Unfortunately, the FPA camera did not give optical power readings and instead only specifies relative optical intensity in counts.)
(86) Imaging with a WFOV Meta-Lens
(87)
(88)
(89) Beam/Image Projection
(90) When light emitter arrays are positioned at the focal plane, the WFOV meta-lens system can be used for beam or image projection with a large projection angle. In
(91) The WFOV meta-lens 100 can be used for projecting an array of beams generated from the emitter array 140 towards a wide range of angles, for applications such as 3D sensing, detection, ranging, communication, etc. The WFOV meta-lens 100 can also be used to project images generated from the emitter array 140 (e.g., a micro-display) towards a wide range of angles, for applications such as display, holography, AR/VR, etc. Some examples are explained below.
(92) 3D Sensing with WFOV Meta-Lenses
(93) Existing 3D depth sensors based on structured light (SL), time-of-flight (TOF), or active stereoscopic technologies are constrained by a small FOV (typically less than 70) and resolution usually limited to around 1,0001,000 resolvable spots or angles. A different yet related technology is vision-based simultaneous localization and mapping (V-SLAM). Intel's RealSense Tracking Camera T265 uses V-SLAM to provide an impressive stitched FOV up to 163, although it has two fisheye imaging units and suffers from lower resolution due to the inferior imaging quality of the fisheye lenses compared to standard camera optics, especially at large field angles.
(94)
(95) The 3D sensor 1100 includes a pattern projecting module 1110 with the first WFOV meta-lens 1112 and a light emitter array 1114 (e.g., a micro-LED or vertical-cavity surface-emitting laser (VCSEL) array) and a camera module 1120 with the second WFOV meta-lens 1122 and photodetector array 1124. (For stereoscopic sensing, the light emitter array 1114 can be replaced by a second photodetector array, turning the patterning projecting module 1110 into another camera module.) The pattern projecting module 1110 and camera module 1120 are mounted on a substrate 1102, which may be a flexible or rigid piece of plastic, glass, or other suitable material. The entire sensor 1100 may be less than 7 mm thick (e.g., 3 mm or thinner) and can be made with commercial light emitter and photodetector arrays.
(96) The pattern projecting module 1110 projects patterns, such as arrays of dots or stripes, into free-space with a large angular range 1111 (e.g., 120, 130, 140, 150, 160, 170, or larger) and onto an object 1101 (e.g., a hand or fingertip). The image of the object 1101 as illuminated by the pattern projector 1110 is captured by the camera module 1120 over an equally wide and overlapping FOV 1121. A processor (not shown) coupled to the pattern projecting module 1110 and the camera module 1120 analyzes the image captured by the camera module 1120 to generate 3D information about the object 1101.
(97) The ultra-wide beam projection and detection angles 1111 and 1121 allow 3D sensing over large spatial and angular ranges. Diffraction-limited performance of the meta-lenses 1112 and 1122 allows high-quality pattern generation and high-resolution imaging. The simple optical module configuration facilitates tight integration into mobile devices, such as smartphones and tablets, and tolerance to assembly misalignment. The fields-of-view for structured light projection and imaging can be close to 180, with diffraction-limited spatial resolution, to provide precise near-surface 3D sensing/imaging.
(98) For example, a sensor 1100 using the NIR meta-lenses 600 shown in
(99) The sensor 1100 in
(100) The meta-surface on the aperture side of the meta-lens 1112 can also contain the aperture by modulating the distribution of its phase, amplitude, spectral, and/or polarization responses to form an effective aperture. For example, this aperture-side meta-surface can modulate the phase distribution of the incident light so that light within the effective aperture region is transmitted or further modulated to propagate towards the backside meta-surface while the incident light outside the aperture is scattered or deflected away from the region for WFOV imaging or sensing. The meta-surface region outside the effective aperture can also be designed to have minimal transmitted power (e.g., by engineering reflection or absorption properties of the meta-surface). The meta-surface region outside the effective aperture can also be designed to possess spectral or polarization filtering properties to block light with certain wavelengths or polarizations.
(101) Compared to state-of-the-art 3D sensing, 3D sensing with WFOV meta-lenses has many advantages: (1) it is universally applicable to SL, TOF, and active stereoscopic 3D depth sensing; (2) it simultaneously solves the FOV limitations associated with illumination and light receiving (camera) optics; (3) the diffraction-limited imaging and image projection capabilities of the meta-surface optics enable fine spatial resolution over the entire FOV, free of aberrations; (4) since the optics are designed to operate at a single NIR wavelength (e.g., for VCSEL illumination) or over a narrowband (e.g., for micro-LED array illumination), very high (e.g., near-unity) optical efficiency can be achieved; (5) an ultra-compact form factor and minimum element count; and (6) as different parts of the meta-surface interact with light at different AOIs, the meta-atoms can be locally configured to reject ambient light outside the operation wavelength. Unlike traditional optical filters whose passband shifts spectrally with AOI and therefore cannot be used over a large angular range, the meta-surface's unique ultra-wide-angle optical filtering capability can significantly boost signal-to-noise ratio (SNR) and dynamic range.
(102) The high-resolution, panoramic 3D sensor 1100 in
(103) WFOV Meta-Lenses for Augmented Reality (AR) and Virtual Reality (VR)
(104) Micro-displays for augmented and virtual reality applications are attracting significant research and development efforts nowadays due to their broad implementation spaces in next generation display technologies. In terms of optical architectures, AR/VR displays can be categorized into two main groups: bulk-optic-based and waveguide based-systems. Waveguide-based systems are of growing interests in recent years due to their compact form factor and the ease of integration with eyeglasses and other devices. State-of-the-art waveguide-based displays are still limited in resolution and FOV. For example, the FOVs of existing commercial waveguide AR/VR systems are typically less than 4545, far less than the range of human vision. In order to improve the display quality, conventional optical engines use complex and multi-element optical systems, which increase system size and weight.
(105)
(106) Assuming an aperture size of 2 mm, owing to its aberration free imaging performance, the angular resolution of the WFOV meta-surface projection optic 1214 is 0.34 mRad at a wavelength of 550 nm, very close to the resolution of a human eye (e.g., about 0.3 mRad). Given a 180 FOV, this results in a resolution of over 90009000 resolvable spots at each RGB wavelength and an approximately 9 mm9 mm FOV on the micro-display. The total thickness of the meta-optical system is only a few millimeters.
(107) Meta-Lenses with Meta-Surface Apertures
(108)
(109)
CONCLUSION
(110) While various inventive embodiments have been described and illustrated herein, those of ordinary skill in the art will readily envision a variety of other means and/or structures for performing the function and/or obtaining the results and/or one or more of the advantages described herein, and each of such variations and/or modifications is deemed to be within the scope of the inventive embodiments described herein. More generally, those skilled in the art will readily appreciate that all parameters, dimensions, materials, and configurations described herein are meant to be exemplary and that the actual parameters, dimensions, materials, and/or configurations will depend upon the specific application or applications for which the inventive teachings is/are used. Those skilled in the art will recognize or be able to ascertain, using no more than routine experimentation, many equivalents to the specific inventive embodiments described herein. It is, therefore, to be understood that the foregoing embodiments are presented by way of example only and that, within the scope of the appended claims and equivalents thereto, inventive embodiments may be practiced otherwise than as specifically described and claimed. Inventive embodiments of the present disclosure are directed to each individual feature, system, article, material, kit, and/or method described herein. In addition, any combination of two or more such features, systems, articles, materials, kits, and/or methods, if such features, systems, articles, materials, kits, and/or methods are not mutually inconsistent, is included within the inventive scope of the present disclosure.
(111) Also, various inventive concepts may be embodied as one or more methods, of which an example has been provided. The acts performed as part of the method may be ordered in any suitable way. Accordingly, embodiments may be constructed in which acts are performed in an order different than illustrated, which may include performing some acts simultaneously, even though shown as sequential acts in illustrative embodiments.
(112) All definitions, as defined and used herein, should be understood to control over dictionary definitions, definitions in documents incorporated by reference, and/or ordinary meanings of the defined terms.
(113) The indefinite articles a and an, as used herein in the specification and in the claims, unless clearly indicated to the contrary, should be understood to mean at least one.
(114) The phrase and/or, as used herein in the specification and in the claims, should be understood to mean either or both of the elements so conjoined, i.e., elements that are conjunctively present in some cases and disjunctively present in other cases. Multiple elements listed with and/or should be construed in the same fashion, i.e., one or more of the elements so conjoined. Other elements may optionally be present other than the elements specifically identified by the and/or clause, whether related or unrelated to those elements specifically identified. Thus, as a non-limiting example, a reference to A and/or B, when used in conjunction with open-ended language such as comprising can refer, in one embodiment, to A only (optionally including elements other than B); in another embodiment, to B only (optionally including elements other than A); in yet another embodiment, to both A and B (optionally including other elements); etc.
(115) As used herein in the specification and in the claims, or should be understood to have the same meaning as and/or as defined above. For example, when separating items in a list, or or and/or shall be interpreted as being inclusive, i.e., the inclusion of at least one, but also including more than one, of a number or list of elements, and, optionally, additional unlisted items. Only terms clearly indicated to the contrary, such as only one of or exactly one of, or, when used in the claims, consisting of, will refer to the inclusion of exactly one element of a number or list of elements. In general, the term or as used herein shall only be interpreted as indicating exclusive alternatives (i.e. one or the other but not both) when preceded by terms of exclusivity, such as either, one of only one of or exactly one of. Consisting essentially of, when used in the claims, shall have its ordinary meaning as used in the field of patent law.
(116) As used herein in the specification and in the claims, the phrase at least one, in reference to a list of one or more elements, should be understood to mean at least one element selected from any one or more of the elements in the list of elements, but not necessarily including at least one of each and every element specifically listed within the list of elements and not excluding any combinations of elements in the list of elements. This definition also allows that elements may optionally be present other than the elements specifically identified within the list of elements to which the phrase at least one refers, whether related or unrelated to those elements specifically identified. Thus, as a non-limiting example, at least one of A and B (or, equivalently, at least one of A or B, or, equivalently at least one of A and/or B) can refer, in one embodiment, to at least one, optionally including more than one, A, with no B present (and optionally including elements other than B); in another embodiment, to at least one, optionally including more than one, B, with no A present (and optionally including elements other than A); in yet another embodiment, to at least one, optionally including more than one, A, and at least one, optionally including more than one, B (and optionally including other elements); etc.
(117) In the claims, as well as in the specification above, all transitional phrases such as comprising, including, carrying, having, containing, involving, holding, composed of, and the like are to be understood to be open-ended, i.e., to mean including but not limited to. Only the transitional phrases consisting of and consisting essentially of shall be closed or semi-closed transitional phrases, respectively, as set forth in the United States Patent Office Manual of Patent Examining Procedures, Section 2111.03.