NARROW BANDPASS IMAGING LENS
20200081190 ยท 2020-03-12
Inventors
Cpc classification
G02B6/29364
PHYSICS
G02B6/2938
PHYSICS
G02B6/2937
PHYSICS
G01S17/894
PHYSICS
H04N13/254
ELECTRICITY
International classification
Abstract
For imaging objects in a field-of-view, a compound lens includes a primary lens, a secondary lens, and a tertiary lens. A combination of the primary lens, the secondary lens, and the tertiary lens has an F-number of not more than 1.25, a field-of-view of at least 70 degrees, and images objects from the field-of-view on a sensor.
Claims
1. A compound lens comprising: a primary lens; a secondary lens; and a tertiary lens, wherein a combination of the primary lens, the secondary lens, and the tertiary lens has an F-number of not more than 1.25, a field-of-view of at least 70 degrees, and images objects from the field-of-view on a sensor.
2. The compound lens of claim 1, the lens further comprising a bandpass filter disposed at a first light path portion, wherein the compound lens with the bandpass filter transmits at least 85 percent of a specified wavelength of the input light to the sensor.
3. The compound lens of claim 2, wherein the bandpass filter is disposed at the first light path portion behind the secondary lens and before the tertiary lens relative to the input light.
4. The compound lens of claim 2, wherein the bandpass filter is disposed at the first light path portion behind the tertiary lens relative to the input light.
5. The compound lens of claim 2, wherein the bandpass filter is disposed at the first light path portion before the primary lens relative to the input light.
6. The compound lens of claim 2, wherein the bandpass filter has a Full Width at Half Maximum (FWHM) in the range of 2-30 nanometers (nm) about the specified wavelength and a roll off slope of not less than 10 decibels (dB)/nm.
7. The compound lens of claim 2, wherein the bandpass filter has a spectral bandpass centered at the specified wavelength selected from the group consisting of 840-860 nm, 930-950 nm, and 800 to 1000 nm.
8. The compound lens of claim 2, the bandpass filter further comprising a stop at the bandpass filter with a size in the range of 2.5 to 4.0 millimeter (mm).
9. The compound lens of claim 8, the lens further comprising a stop disposed behind the secondary lens relative to the input light, wherein a half-angle of a cone of incidence is not greater than 8 degrees through the stop.
10. The compound lens of claim 1, wherein at least one of the primary lens, the secondary lens, and the tertiary lens is aspherical on at least one surface.
11. The compound lens of claim 10, wherein the tertiary lens is aspherical on at least one surface and functions as a field flattener.
12. The compound lens of claim 10, wherein the tertiary lens is aspherical on the surface adjacent to the sensor and functions as a field flattener.
13. The compound lens of claim 1, wherein the compound lens has a Modulus of Optical Transfer Function (MTF) of greater than 50 percent at a spatial frequency of at least 150 line pairs (lp)/mm.
14. The compound lens of claim 1, wherein a relative intensity of the lens is substantially uniform, varying by at most 10 percent over an entire field-of-view of the lens and the total transmission of the lens is at least 85 percent.
15. The compound lens of claim 1, the apparatus further comprising the sensor, wherein the sensor has a diagonal dimension of 1.8 mm.
16. The compound lens of claim 1, wherein the sensor is corrected by at least one of a lookup table and an algorithmic transformation.
17. The lens of claim 1, wherein the primary lens, the secondary lens, and the tertiary lens are formed of moldable glass and the lens has the following prescription: TABLE-US-00005 Aperture Element Power (diopter) Index Abbe number Thickness (mm) (mm) Primary 460 to 375 1.85-1.95 20.86-20.885 1.05-1.15 3.6-3.8 Air gap 0 1.0003 NA 3.6-3.85 Secondary 161 to 171 1.79-1.87 40.78-41.0 1.5-2.35 4.3-4.5 Air gap 0 1.0003 NA 0.14-0.20 Stop 0 1.0003 NA 0 2.5-4.0 Air gap 0 1.0003 NA 0.12-0.60 Tertiary 257 to 263 1.67-1.69 31.06-31.19 2.35-2.49 3.8-4.0 (Aspherical) Air gap 0 1.0003 NA 1.1-2.6
18. The compound lens of claim 1, wherein the temperature of the compound lens is between 40 Celsius (C) degrees and 75C degrees.
19. A system comprising: a light source that generates pulsed light; a sensor that detects the pulsed light; and a compound lens comprising: a primary lens; a secondary lens; and a tertiary lens, wherein a combination of the primary lens, the secondary lens, and the tertiary lens has an F-number of not more than 1.25, a field-of-view of at least 70 degrees, and images objects from the field-of-view on the sensor.
20. A method comprising: imaging objects from a field-of-view of at least 70 degrees with an F-number of not more than 1.25; collimating incident rays, wherein a cone half angle of the collimated input light is less than 10 degrees; transmitting active illumination of the collimated input light at a specified wavelength with Full Width at Half Maximum (FWHM) in the range of 2-30 nanometers (nm) about the specified wavelength and a roll off slope of not less than 10 decibels (dB)/nm; forming an image from the filtered input light on a sensor; and detecting the image with the sensor.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0005] In order that the advantages of the embodiments of the invention will be readily understood, a more particular description of the embodiments briefly described above will be rendered by reference to specific embodiments that are illustrated in the appended drawings. Understanding that these drawings depict only some embodiments and are not therefore to be considered to be limiting of scope, the embodiments will be described and explained with additional specificity and detail through the use of the accompanying drawings, in which:
[0006]
[0007]
[0008]
[0009]
[0010]
[0011]
[0012]
[0013]
[0014]
[0015]
[0016]
[0017]
[0018]
[0019] FIG. 3D is a front view drawing of a sensor according to an embodiment;
[0020]
[0021]
[0022]
[0023]
[0024]
[0025]
DETAILED DESCRIPTION
[0026] Reference throughout this specification to one embodiment, an embodiment, or similar language means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment. Thus, appearances of the phrases in one embodiment, in an embodiment, and similar language throughout this specification may, but do not necessarily, all refer to the same embodiment, but mean one or more but not all embodiments unless expressly specified otherwise. The terms including, comprising, having, and variations thereof mean including but not limited to unless expressly specified otherwise. An enumerated listing of items does not imply that any or all of the items are mutually exclusive and/or mutually inclusive, unless expressly specified otherwise. The terms a, an, and the also refer to one or more unless expressly specified otherwise.
[0027] Furthermore, the described features, advantages, and characteristics of the embodiments may be combined in any suitable manner. One skilled in the relevant art will recognize that the embodiments may be practiced without one or more of the specific features or advantages of a particular embodiment. In other instances, additional features and advantages may be recognized in certain embodiments that may not be present in all embodiments.
[0028] These features and advantages of the embodiments will become more fully apparent from the following description and appended claims, or may be learned by the practice of embodiments as set forth hereinafter. As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method, and/or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a circuit, module, or system. Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having program code embodied thereon.
[0029] Many of the functional units described in this specification have been labeled as modules, in order to more particularly emphasize their implementation independence. For example, a module may be implemented as a hardware circuit comprising custom VLSI circuits or gate arrays, off-the-shelf semiconductors such as logic chips, transistors, or other discrete components. A module may also be implemented in programmable hardware devices such as field programmable gate arrays, programmable array logic, programmable logic devices or the like.
[0030] Modules may also be implemented in software for execution by various types of processors. An identified module of program code may, for instance, comprise one or more physical or logical blocks of computer instructions which may, for instance, be organized as an object, procedure, or function. Nevertheless, the executables of an identified module need not be physically located together, but may comprise disparate instructions stored in different locations which, when joined logically together, comprise the module and achieve the stated purpose for the module.
[0031] Indeed, a module of program code may be a single instruction, or many instructions, and may even be distributed over several different code segments, among different programs, and across several memory devices. Similarly, operational data may be identified and illustrated herein within modules, and may be embodied in any suitable form and organized within any suitable type of data structure. The operational data may be collected as a single data set, or may be distributed over different locations including over different storage devices, and may exist, at least partially, merely as electronic signals on a system or network. Where a module or portions of a module are implemented in software, the program code may be stored and/or propagated on in one or more computer readable medium(s).
[0032] The computer readable medium may be a tangible computer readable storage medium storing the program code. The computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, holographic, micromechanical, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing.
[0033] More specific examples of the computer readable storage medium may include but are not limited to a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a portable compact disc read-only memory (CD-ROM), a digital versatile disc (DVD), an optical storage device, a magnetic storage device, a holographic storage medium, a micromechanical storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, and/or store program code for use by and/or in connection with an instruction execution system, apparatus, or device.
[0034] The computer readable medium may also be a computer readable signal medium. A computer readable signal medium may include a propagated data signal with program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electrical, electro-magnetic, magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport program code for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable signal medium may be transmitted using any appropriate medium, including but not limited to wire-line, optical fiber, Radio Frequency (RF), or the like, or any suitable combination of the foregoing
[0035] In one embodiment, the computer readable medium may comprise a combination of one or more computer readable storage mediums and one or more computer readable signal mediums. For example, program code may be both propagated as an electro-magnetic signal through a fiber optic cable for execution by a processor and stored on RAM storage device for execution by the processor.
[0036] Program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Python, Java, JavaScript, Smalltalk, C++, PHP or the like and conventional procedural programming languages, such as the C programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider). The computer program product may be shared, simultaneously serving multiple customers in a flexible, automated fashion.
[0037] The computer program product may be integrated into a client, server and network environment by providing for the computer program product to coexist with applications, operating systems and network operating systems software and then installing the computer program product on the clients and servers in the environment where the computer program product will function. In one embodiment software is identified on the clients and servers including the network operating system where the computer program product will be deployed that are required by the computer program product or that work in conjunction with the computer program product. This includes the network operating system that is software that enhances a basic operating system by adding networking features.
[0038] Furthermore, the described features, structures, or characteristics of the embodiments may be combined in any suitable manner. In the following description, numerous specific details are provided, such as examples of programming, software modules, user selections, network transactions, database queries, database structures, hardware modules, hardware circuits, hardware chips, etc., to provide a thorough understanding of embodiments. One skilled in the relevant art will recognize, however, that embodiments may be practiced without one or more of the specific details, or with other methods, components, materials, and so forth. In other instances, well-known structures, materials, or operations are not shown or described in detail to avoid obscuring aspects of an embodiment.
[0039] Aspects of the embodiments are described below with reference to schematic flowchart diagrams and/or schematic block diagrams of methods, apparatuses, systems, and computer program products according to embodiments of the invention. It will be understood that each block of the schematic flowchart diagrams and/or schematic block diagrams, and combinations of blocks in the schematic flowchart diagrams and/or schematic block diagrams, can be implemented by program code. The program code may be provided to a processor of a general purpose computer, special purpose computer, sequencer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the schematic flowchart diagrams and/or schematic block diagrams block or blocks.
[0040] The program code may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the schematic flowchart diagrams and/or schematic block diagrams block or blocks.
[0041] The program code may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the program code which executed on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
[0042] The schematic flowchart diagrams and/or schematic block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of apparatuses, systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the schematic flowchart diagrams and/or schematic block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions of the program code for implementing the specified logical function(s).
[0043] It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the Figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. Other steps and methods may be conceived that are equivalent in function, logic, or effect to one or more blocks, or portions thereof, of the illustrated Figures.
[0044] Although various arrow types and line types may be employed in the flowchart and/or block diagrams, they are understood not to limit the scope of the corresponding embodiments. Indeed, some arrows or other connectors may be used to indicate only the logical flow of the depicted embodiment. For instance, an arrow may indicate a waiting or monitoring period of unspecified duration between enumerated steps of the depicted embodiment. It will also be noted that each block of the block diagrams and/or flowchart diagrams, and combinations of blocks in the block diagrams and/or flowchart diagrams, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and program code.
[0045] The schematic flowchart diagrams and/or schematic block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations. It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the Figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. Although various arrow types and line types may be employed in the flowchart and/or block diagrams, they are understood not to limit the scope of the corresponding embodiments. Indeed, some arrows or other connectors may be used to indicate only an exemplary logical flow of the depicted embodiment.
[0046] The description of elements in each figure may refer to elements of proceeding figures. Like numbers refer to like elements in all figures, including alternate embodiments of like elements.
[0047] Lenses for a wide field-of-view are typically compound lenses. The need to compensate and balance optical aberrations, which increase at the edges of the field-of-view, is fulfilled by the combinations of multiple elements having dioptric power and dissimilar refractive indices and Abbe numbers. It follows that the number of lens elements grows with the angle-of-view, or the field-of-view. Starting with the Cooke lens, a type of triplet lens with air spaced elements, that comprises a negative flint glass element in the center with a positive crown glass element on each side, increasing the field beyond a certain angle, say 40 degrees, requires the addition of a fourth element, rendering the new assembly a Tessar lens. Though both the Cooke and Tessar lenses were superseded by more advanced lenses, they made a comeback in recent years in devices such as mobile-phone cameras.
[0048] Lenses used in mobile-phone cameras consist typically of four to five powered elements. Because such lenses rely on abundant background illumination they do not require low F-number and are designed for an F-number of 4 or greater. Since the background does not present a problem they use broadband optical filters with gradual roll-on and roll-off curves.
[0049] In cameras operating under low illumination conditions the F-number of the lens, a parameter that determines the entrance pupil of the lens, is of paramount importance since the lens ability to collect luminous or radiant flux is inversely proportional to the square of the F-number. It follows that cameras with low illumination level require a low F-number. Consistent with this scenario is a case in which background illumination must be minimized and where active illumination is applied to permit imaging of the field-of-view (FOV). In these cases a bandpass filter is disposed in the optical train to minimize such background while transmitting substantially the entire active illumination.
[0050] Because an optimal bandpass filter has a flattop transmission spectrum with steep roll-off curves about the center wavelength and strong blocking outside of the transmission band, the incidence angle of rays on the filters must be kept at low values. To enable this, a quasi-collimated zone is created in the lens where the bandpass filter is disposed. The embodiments provide a fast-aperture lens with an F-number of 1.25 made of three powered lens elements, for a field-of-view of 70 degrees. A certain embodiment includes the bandpass filter.
[0051]
[0052] In one embodiment, all the powered lens elements including the primary lens 101, the secondary lens 103, and the tertiary lens 105 are made of moldable glass. At least one of the primary lens 101, the secondary lens 103, and the tertiary lens 105 may be aspherical on at least one surface. In a certain embodiment, the tertiary lens 105 is aspherical on at least one surface and functions as a field flattener. In one embodiment, the at least one surface is closest to the sensor 109.
[0053] The combination of the primary lens 101, the secondary lens 103, and the tertiary lens 105 may have an F-number of not more than 1.25, a field-of-view of at least 70 degrees, and focus rays of input light 111 thus forming an image of objects in the field-of-view on the sensor 109. The lens 100 includes a plurality of light path portions 123.
[0054] In one embodiment, the lens length 141 is in the range of 9 to 13 millimeters (mm). In addition, the total track length 143 may be in the range of 14 to 18 mm. In a certain embodiment, the lens length 141 is 11 mm and the total track length 143 is 16 mm. Table 1 illustrates one embodiment of a prescription for the lens 100.
TABLE-US-00001 TABLE 1 Aperture Element Power (diopter) Index Abbe number Thickness (mm) (mm) Primary 460 to 375 1.85-1.95 20.86-20.885 1.05-1.15 3.6-3.8 Air gap 0 1.0003 NA 3.6-3.85 Secondary 161 to 171 1.79-1.87 40.78-41.0 1.5-2.35 4.3-4.5 Air gap 0 1.0003 NA 0.14-0.20 Stop 0 1.0003 NA 0 2.5-4.0 Air gap 0 1.0003 NA 0.12-0.60 Tertiary 257 to 263 1.67-1.69 31.06-31.19 2.35-2.49 3.8-4.0 (Aspherical) Air gap 0 1.0003 NA 1.1-2.6 Sensor
[0055] The sensor 109 may have a diagonal dimension of 1.5 to 2.0 mm. In a certain embodiment, the sensor 109 has a diagonal dimension of 1.8 mm ( 1/14 inch), making the lens 100 compatible with extremely compact sensors 109. The diagonal dimension is illustrated in FIG. 3D. Though offering a smaller resolution, small sensors 109 are advantageous in the amount of light received in comparison with large sensors 109. The sensor 109 may be a semiconductor device such as a charge coupled device (CCD) and/or a complimentary metal-oxide semiconductor (CMOS) device. In an embodiment, the 1.8 mm sensor 109 has a pixel pitch of 10 micrometers, thus a total of 15,000 pixels for a sensor aspect ratio of 3:2, approximately 15,550 for a sensor aspect ratio of 4:3 and approximately 15,800 a sensor aspect ratio of 5:4. Considering that typically the electronic bandwidth of an individual pixel in a CCD sensor 109 is 10 MHz, this determines a sensor bandwidth of 700 Hz, or integration time of 0.5 millisecond (ms) for the sensor 109 with 15,000 pixels. In a certain embodiment, the sensor 109 is of the CMOS type, having typically an electronic bandwidth of 100 kHz per row, or an integration time of about 0.35 ms per 100 rows. The integration time is approximately inversely proportional to the number of pixels.
[0056] In one embodiment, the total transmission of the lens 100 at a specified wavelength of input light 111 is at least 85 percent. In one embodiment, the lens 100 has a MTF of greater than 50 percent at a spatial frequency of at least 150 line pairs (lp)/mm. The MTF of the lens 100 is described in more detail in
[0057]
[0058] The lens 100 with the bandpass filter 107 may transmit at least 85 percent of a specified wavelength of the input light 111 to the sensor 109. In an embodiment, the lens 100 images a field-of-view of at least 70 degrees, with an F-number of not more than 1.25. In addition, the lens 100 exhibits a diffraction limited MTF up to 150 line lp/mm, uniform relative illumination and distortion of up to 18 percent. The total in band transmission of the lens 100 at a specified wavelength of input light 111 may be in the range of 75 percent to 95 percent. In a certain embodiment, the total in band transmission of the bandpass filter 107 at the specified wavelength of input light 111 is 95 percent and the out of band transmission is substantially opaque. As used herein, substantially opaque refers to a transmission of less than 1 percent.
[0059] In one embodiment, the bandpass filter 107 has a Full Width at Half Maximum (FWHM) in the range of 2-30 nm about the specified wavelength and a roll off slope of not less than 10 dB/nm. The bandpass filter 107 may have a spectral bandpass centered at a specified wavelength selected from the group consisting of 840-860 nm, 930-950 nm, and 800-1000 nm. In one embodiment, the compound lens 100 forms an optimally sharp image using light at the wavelengths of in the range of 840-860 nm and/or 930-950 nm. In a certain embodiment, the compound lens 100 forms an optimally sharp image using light at the wavelengths in the range of 800-1000 nm.
[0060] In one embodiment, the bandpass filter 107 has a spectral bandpass centered at 857 nm and a Full Width at Half Maximum (FWHM) of 30 nm. In one example, if the background radiation is that of the sun, casting an integrated irradiance over the entire spectrum of 1 kW/m.sup.2 on an object in the field-of-view, then the background irradiance is reduced by the bandpass filter 107 to approximately 40 watts (W)/meter.sup.2 (m) through the bandpass filter 107. The effective background irradiance may be further reduced proportionally by using a narrower bandpass filter 107. In an exemplary embodiment bandpass filter 107 with a FWHM of 2 nm may be employed. In one embodiment, irradiance included by the active illumination on an object in the field-of-view may exceed 40 W/m.sup.2.
[0061] Conversely, placement of a bandpass filter 107 behind an imaging lens exposes the filter to incidence angles of the entire FOV. This degrades the bandpass filter performance in three ways: shifting the center wavelength to the blue side of the spectrum, reduction of the peak transmission and relaxing the rollon-rolloff slopes. For instance, for a case where the cone half-angle is 45 the effect on the above mentioned bandpass filter 107 would be: a center wavelength shift to 825 nm, a peak transmission drop to 50 percent and broadening of the FWHM to 70 nm.
[0062] In one embodiment, the lens length 141 is in the range of 9 to 13 millimeters (mm). In addition, the total track length 143 may be in the range of 14 to 18 mm. In a certain embodiment, the lens length 141 is 11 mm and the total track length 143 is 16 mm. A prescription of one embodiment of the lens 100 is given in Table 2.
TABLE-US-00002 TABLE 2 Power Abbe Thickness Aperture Element (diopter) Index number (mm) (mm) Primary 455.52 1.89 20.87 1.1 3.7 Air gap 0 1.0003 NA 3.786 Secondary 167.48 1.80 40.84 1.572 4.4 Air gap 0 1.0003 NA 0.15 Stop 0 1.0003 NA 0 3.2 Air gap 0 1.0003 NA 0.15 Filter 0 1.51 64.17 2 4 Air gap 0 1.0003 NA 0.150 Tertiary 262.64 1.68 31.08 2.406 3.9 (Aspherical) Air gap 0 1.0003 NA 2.6 Sensor
[0063]
[0064]
[0065] In one embodiment, the lens length 141 is in the range of 9 to 13 millimeters (mm). In addition, the total track length 143 may be in the range of 14 to 18 mm. In a certain embodiment, the lens length 141 is 11 mm and the total track length 143 is 16 mm. A prescription of one embodiment of the lens 100 is given in Table 3.
TABLE-US-00003 TABLE 3 Power Abbe Thickness Aperture Element (diopter) Index number (mm) (mm) Primary 449.89 1.923 20.883 1.1 3.7 Air gap 0 1.0003 NA 3.550 Secondary 170.81 1.806 40.95 1.778 4.4 Air gap 0 1.0003 NA 0.197 Stop 0 1.0003 NA 0 3.2 Air gap 0 1.0003 NA 0.127 Filter 0 1.51 64.17 2 4 Air gap 0 1.0003 NA 0.190 Tertiary 259.66 1.689 31.185 2.483 3.9 (Aspherical) Air gap 0 1.0003 NA 1.536 Sensor 0 1.5233 54.517 0.55 window Air gap 0 1.0003 NA 1.536 Sensor
[0066]
[0067]
[0068] In one embodiment, the lens length 141 is in the range of 9 to 13 millimeters (mm). In addition, the total track length 143 may be in the range of 14 to 24 mm. In a certain embodiment, the lens length 141 is 11 mm and the total track length 143 is 16 mm. A prescription of one embodiment of the lens 100 in
TABLE-US-00004 TABLE 4 Power Abbe Thickness Aperture Element (diopter) Index number (mm) (mm) Primary 378.25 1.923 20.883 1.1 3.7 Air gap 0 1.0003 NA 4.143 Secondary 161.79 1.806 40.95 2.311 4.4 Air gap 0 1.0003 NA 0.18 Stop 0 1.0003 NA 0 3.2 Air gap 0 1.0003 NA 0.544 Tertiary 258.61 1.689 31.185 2.368 3.9 (Aspherical) Air gap 0 1.0003 NA 1.209 Sensor 0 1.5233 54.517 0.7 window/ filter Air gap 0 1.0003 NA 0.6 Sensor
[0069]
[0070]
[0071] In one embodiment, a light source 145 generates active illumination light 141. The active illumination light 141 may be to illuminate the environment for three-dimensional (3D) photography. The active illumination light 141 may be a pulsed light active illumination light 141. The active illumination light 141 may be a laser active illumination light 141. In a certain embodiment, the active illumination light 141 is a lidar active illumination light 141. The active illumination light 141 may scatter and/or reflect off object 143 in the environment and be received as an active illumination input light 111a by the lens 100. Thus, the active illumination input light 111a may include the active illumination light 141.
[0072] The embodiments filter the background radiation light 111b from the active illumination input light 111a to maintain illumination magnitude exceeding that of the background illumination. As a result, the sensor 109 may detect the active illumination input light 111a for lidar navigation, 3D photography, or the like. The example, the optical system 150 may be incorporated in a lidar system that detects objects 143.
[0073]
[0074]
[0075]
[0076]
[0077]
[0078]
[0079]
[0080]
[0081] The method 500 starts, and in one embodiment, the lens 100 images 505 objects 143 from the environment. The input light 111 may include both background radiation input light 111b and an active illumination input light 111a. The active illumination input light 111a may be at a specified wavelength and/or a specified wavelength range. In one embodiment, the lens 100 receives the input light 111 over field-of-view of at least 70 degrees. Because the F-number of the lens 100 is not more than 1.25, the lens 100 gathers significantly more light than would be gathered by a lens with a greater F-number.
[0082] The lens 100 may collimate 510 the incident rays of the input light 111. The incident rays may be collimating by the combination of one or more of the primary lens 101, the secondary lens 103, and the tertiary lens 105. The cone half angle of the collimated input light 111 in the light path portion 123 may be less than 10 degrees.
[0083] In a certain embodiment, the lens 100 transmits 515 the specified wavelength of the active illumination input light 111a and filters out the background radiation input light 111b so that the active illumination input light 111a may be detected with a greater signal-to-noise ratio. The bandpass filter 107 may transmit at least 85 percent of the specified wavelength of the input light 111 to the sensor 109. In one embodiment, the bandpass filter 107 has a FWHM in the range of 2-30 nm about the specified wavelength and a roll off slope of not less than 10 dB/nm. The bandpass filter 107 may have a spectral bandpass centered at a specified wavelength selected from the group consisting of 840-860 nm, 930-950 nm, and 800-1000 nm.
[0084] The lens 100 further forms 520 an image from the filtered input light 111 on the sensor 109. The sensor 109 may detect 525 the object 143 in the field-of-view of the compound lens 100 at the specified wavelength and/or the specified wavelength range of the active illumination input light 111a and the method 500 ends. The detected active illumination input light 111a may be used to determine the 3-dimensional position in lens coordinates of an object 143 relative to the lens 100.
[0085]
[0086]
[0087]
[0088] Problem/Solution
[0089] Applications such as autonomous navigation and 3D photography require detecting an active illumination input light 111a from background radiation input light 111b. Unfortunately, when the background radiation input light 111b is strong, particularly in direct sunlight, it is difficult to filter the background radiation input light 111b from the desired active illumination input light 111a. Bandpass filters 107 have been used, but because of the wide field-of-view in many applications, the bandpass filters 107 had unacceptably low transmission of desired specified wavelengths and wide acceptance of the background light.
[0090] The embodiments provide a lens 100 that collimates the input light 111 from a wide field-of-view. The collimated input light 111 may be filtered by the bandpass filter 107 to transmit up to 95 percent of the specified wavelength of the active illumination input light 111a while the transmission of the background radiation input light 111b is substantially reduced. As a result, the sensor 109 can detect the active illumination input light 111a. The design of the lens 100 increases the efficiency of the sensor 109.
[0091] This description uses examples to disclose the invention and also to enable any person skilled in the art to practice the invention, including making and using any devices or systems and performing any incorporated methods. The patentable scope of the invention is defined by the claims and may include other examples that occur to those skilled in the art. Such other examples are intended to be within the scope of the claims if they have structural elements that do not differ from the literal language of the claims, or if they include equivalent structural elements with insubstantial differences from the literal language of the claims.