System for actinic inspection of semiconductor masks

11002688 · 2021-05-11

    Inventors

    Cpc classification

    International classification

    Abstract

    An apparatus and method are disclosed for actinic inspection of semiconductor masks intended for extended ultraviolet (EUV) lithography, or similar objects, with feature sizes less than 100 nm. The approach uses a coherent light source with wavelength less than 120 nm. Inside a vacuum system, an optical system directs the light to an object, i.e., the mask or mask blank, and directs the resulting reflected or transmitted light to an imaging sensor. A computational system processes the imaging sensor data to generate phase and amplitude images of the object. The preferred imaging modality, a form of digital holography, produces images of buried structures and phase objects, as well as amplitude or reflectance images, with nanometer resolution less than or equal to the feature size of the mask.

    Claims

    1. An apparatus for imaging an object with resolution less than 100 nm, the apparatus comprising: a coherent light source with wavelength less than 120 nm; an optical system for directing the source light to an object and directing light from the object to an imaging sensor, wherein the optical system divides the source light into a first beam and a second, different beam and directs the first beam to illuminate the object and directs the second beam to interfere with the first beam, and wherein the angle between the two beams is less than 6.75 mrad; and a computational system that processes the imaging sensor data to generate phase and amplitude images of the object.

    2. The apparatus according to claim 1 wherein the coherent light source uses high harmonic generation (HHG) for producing the source light.

    3. The apparatus according to claim 1 wherein the coherent light source uses a free electron laser (FEL) for producing the source light.

    4. The apparatus according to claim 1 wherein the optical system comprises a diffractive optic.

    5. The apparatus according to claim 1 wherein the optical system comprises a compound zone plate optic comprising a transmissive diffractive optic.

    6. The apparatus according to claim 1 wherein the optical system comprises a series of mirrors to direct the second beam.

    7. The apparatus according to claim 1 wherein the object is a mask designed for semiconductor lithography, and further wherein the coherent light source has the same wavelength as is used with the mask for semiconductor lithography.

    8. The apparatus according to claim 1 wherein the imaging sensor comprises an imaging array with pixels 6 microns square or smaller.

    9. An apparatus for imaging an object, the apparatus comprising: a coherent light source; an optical system for directing the source light to an object and directing light from the object to an imaging sensor, wherein the optical system divides the source light into a first beam and a second, different beam and directs the first beam to illuminate the object and directs the second beam to interfere with the first beam, and wherein the two beams have optical aberrations which produce straighter interference fringes than purely spherical waves; and a computational system that processes the imaging sensor data to generate phase and amplitude images of the object.

    10. A method for imaging an object with resolution less than 100 nm, the method comprising: providing coherent light with wavelength less than 120 nm; directing the light to an object and directing light from the object to an imaging sensor, wherein an optical system divides the source light into a first beam and a second, different beam and directs the first beam to illuminate the object and directs the second beam to interfere with the first beam, and wherein the angle between the two beams is less than 6.75 mrad; and processing the imaging sensor data to generate phase and amplitude images of the object.

    11. The method according to claim 10 wherein the coherent light is produced using high harmonic generation (HHG).

    12. The method according to claim 10 wherein the coherent light is produced using a free electron laser (FEL).

    13. The method according to claim 10 wherein the optical system comprises a diffractive optic.

    14. The method according to claim 10 wherein the optical system comprises a compound zone plate optic comprising a transmissive diffractive optic.

    15. The method according to claim 10 wherein a series of mirrors are used to direct the second beam.

    16. The method according to claim 10 wherein the object is a mask designed for semiconductor lithography, and further wherein the coherent light source has the same wavelength as is used with the mask for semiconductor lithography.

    17. The method according to claim 10 wherein the imaging sensor comprises an imaging array with pixels 6 microns square or smaller.

    18. The method according to claim 10 wherein the processing comprises a Fourier masking algorithm to recover the propagated complex mask pattern.

    19. The method according to claim 10 wherein the processing comprises an algorithm wherein a diffraction propagation algorithm is applied to the propagated complex mask pattern to generate phase and amplitude images of the object.

    20. A method for imaging an object, the method comprising: providing coherent light; directing the light to an object and directing light from the object to an imaging sensor, wherein an optical system divides the source light into a first beam and a second, different beam and directs the first beam to illuminate the object and directs the second beam to interfere with the first beam, and wherein the two beams have optical aberrations which produce straighter interference fringes than purely spherical waves; and processing the imaging sensor data to generate phase and amplitude images of the object.

    21. The apparatus according to claim 9 wherein the optical system comprises a diffractive optic.

    22. The apparatus according to claim 9 wherein the optical system comprises a compound zone plate optic comprising a transmissive diffractive optic.

    23. The apparatus according to claim 9 wherein the optical system comprises a series of mirrors to direct the second beam.

    24. The apparatus according to claim 9 wherein the object is a mask designed for semiconductor lithography, and further wherein the coherent light source has the same wavelength as is used with the mask for semiconductor lithography.

    25. The method according to claim 20 wherein the optical system comprises a diffractive optic.

    26. The method according to claim 20 wherein the optical system comprises a compound zone plate optic comprising a transmissive diffractive optic.

    27. The method according to claim 20 wherein a series of mirrors are used to direct the second beam.

    28. The method according to claim 20 wherein the object is a mask designed for semiconductor lithography, and further wherein the coherent light source has the same wavelength as is used with the mask for semiconductor lithography.

    29. The method according to claim 20 wherein the processing comprises a Fourier masking algorithm to recover the propagated complex mask pattern.

    30. The method according to claim 20 wherein the processing comprises an algorithm wherein a diffraction propagation algorithm is applied to the propagated complex mask pattern to generate phase and amplitude images of the object.

    Description

    BRIEF DESCRIPTION OF THE DRAWINGS

    (1) The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.

    (2) The foregoing features of the invention will be more readily understood by reference to the following detailed description, taken with reference to the accompanying drawings, wherein:

    (3) FIG. 1 is a graph of the number of exposures required to image an entire standard EUV mask vs the resolution for 2K×2K images per exposure;

    (4) FIG. 2 is a schematic view of a novel optical system formed in accordance with the present invention;

    (5) FIG. 2A is a schematic view of an alternate optical system formed in accordance with the present invention;

    (6) FIG. 3 shows exemplary plots of on-axis, off-axis, and compound Fresnel zone plates (FZPs) along with resulting images of focused beams at the focal distance and at 0.97× the focal distance;

    (7) FIG. 4 is an image of a portion of a FZP designed for EUV imaging;

    (8) FIG. 5 is an exemplary timing diagram showing the sequence of exposure, readout, linear position of the mask and its acceleration when obtaining a sequence of images across the mask;

    (9) FIG. 6 shows intensity fringes from two interfering spherical waves with and without compensating aberrations and their difference, and the corresponding spatial frequency components;

    (10) FIG. 7 shows the USAF resolution target used as the mask pattern in the imaging simulation;

    (11) FIG. 8 shows logarithmic plots of the Fourier transform of the intensity of two interfering point sources with and without compensating aberrations;

    (12) FIG. 9 shows simulation images of the propagated mask pattern and its interference with a reference beam;

    (13) FIG. 10 shows various simulation outputs of the Fourier Masking (FM) algorithm and back propagation estimates of the mask pattern with and without compensating aberrations; and

    (14) FIG. 11 shows the phase of the back-propagated estimate of the mask pattern.

    DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

    (15) In accordance with preferred embodiments of the present invention, and looking now at FIG. 2, there is provided a novel system comprising a source of coherent electromagnetic radiation 100, typically EUV or soft x-ray, an optical system 200, an object (mask) 300 to be imaged, a motion system 400 for translating the object (mask) 300 relative to the optical system, an imaging sensor 500, an algorithm 600 implemented by a computational system 700, all of which are used to produce an estimate 800 of object (mask) 300 and present the estimate on a display 900 with resolution less than 100 nm.

    (16) In order to easily measure the complex reflectivity (amplitude and phase) of the mask, the imaging modality is a coherent imaging system. In order to obtain the mask image deterministically, an imaging method similar to imaging interferometry or digital holography is employed. These methods rely on measuring the interference of a coherent beam reflected or scattered by the object with a coherent reference beam.

    Source of Coherent Electromagnetic Radiation 100

    (17) In order to easily image features that are phase defects, light source 100 must be coherent. Previously, that required using synchrotron radiation, which is produced by an accelerating beam of charged particles. More recently, researchers have found that it is possible to generate coherent beams of extreme ultraviolet (EUV, 120 nm−10 nm) and soft x-ray (SXR, 10 nm−0.1 nm) radiation using a process termed high harmonic generation (HHG). In this process, a pulse from an ultrafast laser, typically with a pulse length measured in femtoseconds (fs), is incident on a gas. The intense laser pulse (“the pump laser”) partially ionizes the gas and excites a high harmonic of the laser frequency (“the signal beam”) when the ionized electrons are accelerated and slam into the nucleus (see Aria J. Paul, Ph.D. Dissertation: “Coherent EUV Light from High-Order Harmonic Generation: Enhancement and Applications to Lensless Diffractive Imaging”, University of Colorado, 2007).

    (18) While this is a useful technique, the power of the resulting coherent EUV beam is small. A variety of factors—the pump wavelength, the laser pulse energy and pulse length, the noble gas chosen, the gas pressure, and the waveguide geometry—affect the efficiency of the HHG process. With optimal choices, EUV and SXR conversion efficiencies of 10.sup.−6 to 10.sup.−3 have been achieved with ultrafast lasers having average powers from 1 W-10 W. The resulting short wavelength beam is nearly collimated and has good beam quality and spatial coherence. These microwatt to milliwatt beams are bright enough for many imaging applications. However, the power which has been achieved at the EUV 13.5 nm wavelength is at the low end of this range (see KMLabs XUUS product, https://www.kmlabs.com/product/xuus/). As discussed below in this specification, that power (i.e., the power achieved at the EUV 13.5 nm wavelength) may be insufficient to scan the entire mask in a reasonable period while achieving sufficient signal-to-noise ratio (SNR). This could limit the application of systems using HHG coherent sources to tasks requiring less throughput, e.g., defect review. However, alternate sources of coherent EUV beams are available.

    (19) It is possible to generate a coherent EUV beam using a compact FEL. EUV beams with 20 mW of power have been demonstrated by Ackermann et al., at a large facility (see Operation of a free-electron laser from the extreme ultraviolet to the water window, W. Ackermann et al., Nature Photonics 1, 336-342 (2007), Published online: 1 Jun. 2007 I doi:10.1038/nphoton.2007.76). A variety of approaches are being tested for developing compact FELs that should yield comparable power outputs. With no driving input, an FEL will lase when spontaneous excitation begins stimulated emission in a process termed self-amplified spontaneous emission (SASE). Beams with greater temporal coherence and higher output efficiency can be generated when an input beam seeds the stimulated emission. Regarding coherence, Allaria et al. demonstrated a seeded EUV FEL beam with fractional bandwidth of 0.16% (ΔE/E) (see Allaria, E et al. Highly coherent and stable pulses from the FERMI seeded free-electron laser in the extreme ultraviolet, Nature Photonics 6(10). p. 699-704, 2012). It will be seen below that coherent power in the milliwatt range is sufficient for the imaging system of the present invention. Thus a sufficiently high powered and highly coherent EUV beam can be engineered using a compact FEL.

    (20) Thus, in accordance with the present invention, light source 100 is configured to provide a coherent EUV beam, and in one preferred form of the present invention, light source 100 comprises a compact FEL.

    Optical System 200

    (21) For imaging amplitude and phase objects, a coherent imaging approach with interfering beams, such as imaging interferometry or off-axis holography, is preferred. The optics used, and the various parameters of this approach, must be carefully engineered to optimize the information collection.

    (22) There are no transmissive optical materials at the EUV actinic wavelength, 13.5 nm, and even air at atmospheric pressure has an absorption length measured in mm. Therefore, reflective optics and a reflective mask are used exclusively for projecting the mask onto the wafer and the mask and optics must be in vacuum. This applies equally to an exposure tool and to any actinic inspection tools. By design, the optical axis is 6° from normal for first generation EUV lithography tools so that the beam reflected off the mask does not intersect the illuminator optics. This is the limiting angle for the edge ray of the illuminating beam in the plane of incidence. For a rotationally symmetric beam, the 6° angle is roughly equivalent to a numerical aperture (NA) 0.10 beam incident on the mask.

    (23) First generation EUV lithography tools image the mask onto the wafer with 4× reduction, and newer tools are under development with anamorphic optics that have 4× reduction in one direction and 8× reduction in the orthogonal direction. The demagnification of the mask yields a beam with larger NA at the wafer. The first generation tools (e.g., the NXE 3100 by ASML of Veldhoven, The Netherlands) have a 0.25 NA beam at the wafer and the current generation tools (e.g., the NXE 3400B by ASML) use a 0.33 NA beam. Future generations with 0.55 NA and anamorphic beams are in development. Proposed systems with anamorphic beams take advantage of the fact that the beam angle in the direction orthogonal to the plane of incidence may be larger, thereby increasing the resolution in that direction.

    (24) Assuming the optics are diffraction limited, the finest possible resolution of the system is given by

    (25) δ R = k R λ NA ( 1 )
    where k.sub.R is a proportionality constant that depends on details of the imaging system and the definition of resolution, and λ is the wavelength.

    (26) In the mask plane, the tolerance for the mask position along the optical axis (similar to the depth of focus, DOF) is given by

    (27) δ DOF = k DOF λ NA 2 ( 2 )
    for a displacement that leaves the resolution essentially unchanged, again assuming diffraction limited optics, where k.sub.DOF is a proportionality constant that depends on details of the imaging system and the definition of depth of focus.

    (28) An optimally designed system starts with the relationship between the number of pixels of the imaging sensor 500 (see FIG. 2) and the number of pixels in the reconstructed image. Assuming reasonably uniform distribution (on average) of incident radiation on imaging sensor 500, one would want 2× as many pixels at the imaging sensor for an imaging method that recovers amplitude and phase. Another factor of 2 or larger is necessary to break the symmetry of an object and its complex conjugate. Thus the target region on the object typically has, at most, ¼ of the number of pixels of the imaging sensor.

    (29) Note that this relationship holds when the pixel scale of the imaging sensor 500 matches the scale of features of the radiation pattern incident upon it. Namely, the radiation incident on the imaging sensor should roughly cover the imaging sensor and have its significant scales resolved by the pixel scale. This requires that the object illumination should correspond to the resolvable area. If the imaging sensor is an N×N pixel array, then the resolvable area is roughly an M×M square where M=N/2×(kλ/NA). For a λ=13.5 nm EUV beam with k=0.5, NA=0.125, and a 4K×4K imaging sensor, the resolvable area is roughly 110 μm square, with the resolution element 54 nm, or 13.5 nm after 4× demagnification at the wafer.

    (30) Implicit in this analysis is that the imaging sensor is sized to capture essentially all of the diffracted light and its pixels are sized accordingly so they properly sample the diffraction pattern. Analyses of the appropriate sizes and sampling for the object, wavelength, imaging sensor, and its pixel pitch that elaborate on these rules of thumb have been outlined in several publications (see (i) Huaying Wang; Dayong Wang; Jianjun Xie; Shiquan Tao; Recording conditions of digital holography. Proc. SPIE 6279, 27th International Congress on High-Speed Photography and Photonics, 62791J (Jan. 11, 2007); doi:10.1117/12.725187; and (ii) Jorge Garcia-Sucerquia, Wenbo Xu, Stephan K. Jericho, Peter Klages, Manfred H. Jericho, and H. Jürgen Kreuzer, “Digital in-line holographic microscopy,” Appl. Opt. 45, 836-850 (2006) https://doi.org/10.1364/AO.45.000836), which publications are hereby incorporated herein by reference. Their conclusions for a well designed system match the aforementioned relationships but may differ slightly based on the definition of resolution, e.g., the “k” constant in kλ/NA.

    (31) In order to match the scale of the radiation incident on the imaging sensor 500 to the size of the imaging sensor, one could set the distance from the object (i.e., mask 300) to the imaging sensor accordingly. However, a more compact system results when the radiation incident on the imaging sensor is an expanding beam, as shown in FIG. 2. In general, an expanding beam can be produced by a positive (focusing) or negative optic that is either transmissive or reflective. However, to simplify the data processing, the illuminated region should be restricted to the resolvable area. For efficient use of the illuminating beam, it should just cover the resolvable area.

    (32) This last consideration makes it preferable to use a transmissive optic to convert the coherent EUV beam to a diverging beam, due to the short focal distance for a reasonably large NA. It is easily appreciated that a reflective optic would likely interfere with the mask since it must have a focal length <1 mm given the typical beam diameter and required NA for high resolution. Although conventional transmissive optics are not available for EUV and SXR wavelengths, transmissive diffractive optics are available for EUV and SXR wavelengths and are discussed below.

    (33) In order to generate the two interfering beams (i.e., the first beam 202 which illuminates the object and the second, different beam 204 which interferes with first beam 202 at imaging sensor 500), it is not possible to directly adapt the methods used at longer wavelengths (e.g., DUV wavelengths) for use at EUV wavelengths. The transmissive optics and beamsplitters normally used at longer wavelengths do not exist for 13.5 nm EUV beams. Instead, a novel approach that uses relatively simple optics which are practical for use at 13.5 nm is used. An exemplary optical system is shown in FIG. 2.

    (34) A coherent beam 206 is directed with a planar reflector (fold mirror) R.sub.1 to a transmissive optic, a zone plate 208. A zone plate is the only transmissive element available at the EUV actinic wavelength. This is not an ordinary Fresnel zone plate (FZP) with concentric zones. Instead, it is a compound optic (i.e., a compound zone plate) that forms two off-axis converging (focused) beams. One beam U.sub.0 (i.e., first beam 202) illuminates a portion of the mask (the region being imaged). The second beam U.sub.1 (i.e., second, different beam 204) is a reference beam which is directed by fold mirrors R.sub.2, R.sub.3 so that it does not reflect off the mask. The mirrors R.sub.2, R.sub.3 are arranged so that the two beams U.sub.0 (i.e., first beam 202) and U.sub.1 (i.e., second, different beam 204) have identical path lengths when they overlap and interfere at the detector plane (i.e., at imaging sensor 500). Due to the small wavelength, the angle between the two beams U.sub.0 and U.sub.1 must be small in order to produce fringes that can be resolved by the imaging detector. This is required by the path length requirement
    d sin θ=λ  (3)
    where d is the lateral spacing between adjacent fringe maxima, θ is the angle between the beams, and λ is the wavelength. At the Nyquist sampling limit, d is 2× the detector pixel spacing. For 3 μm pixels, the angle is 2.25 mrad or 0.129 degrees. For 1 μm pixels, which may be close to the practical lower limit that can be fabricated and has sufficient well depth, the angle is 6.75 mrad or 0.387 degrees. The requirement to form the two interfering beams U.sub.0 and U.sub.1 with this small angle presents a challenge for this approach. For expanding beams with nearly spherical wavefronts, the relevant angle is the angle between the two virtual point sources of the spherical wavefronts as viewed from the imaging sensor 500.

    (35) That challenge is reflected in the layout shown in FIG. 2. The mask 300 is positioned in the converging illumination beam U.sub.0 (i.e., first beam 202). In the mask plane, a circular region is illuminated. The beam amplitude is multiplied by P.sub.M, the complex reflectivity of the mask pattern, and becomes U′.sub.0. The illumination beam 202 goes through focus after reflection by the mask 300, as can be seen in the detailed inset A (circle). The reference beam U.sub.1 (i.e., second, different beam 204) is reflected by mirror R2 so that it crosses the illumination beam U.sub.0 and also goes through focus after this reflection. The reference beam U.sub.1 crosses the illumination beam near that beam's minimal width in order to make room for mirror R3 which directs the reference beam U.sub.1 towards the detector (i.e., imaging sensor 500).

    (36) With analogy to digital holography, the first (illuminating) beam U.sub.0 is the object beam which encodes the information about the region being imaged. The second beam U.sub.1 is an off-axis reference beam. At the detector plane, the object beam is a diffracted transformation of the illuminated object with an expanding spherical wavefront. The reference beam is also an expanding spherical wavefront. The two beams U′.sub.0 and U.sub.1 interfere at the detector to form a fringe pattern. If the mask (object) 300 is a perfect planar mirror, the interference fringes are those of two overlapping spherical waves. If the mask (object) 300 has a pattern with varying complex reflectivity (modulus and phase), the fringes are modulated by the object diffraction pattern.

    (37) For these EUV and SXR wavelengths, FZP optics, a type of diffractive optic that typically has two (binary) values, can be fabricated to provide the desired zone plate 208. An FZP (see FIGS. 3 and 4) has alternating concentric rings (zones) of transmissive and opaque material. Since any substrate with sufficient mechanical stability has poor transmission at EUV and SXR wavelengths, FZPs for these wavelengths are constructed like a stencil. Typically made with a thin metal foil, the transmissive zones consist of arc-like perforations (holes) in the foil which follow the circular pattern except for some supporting ribs as is seen in FIG. 4 (see Goldberg et al., “Commissioning an EUV mask microscope for lithography generations reaching 8 nm,” Proc. SPIE 8679 (2013)).

    (38) For an ideal FZP, the spot size is typically quite close to the width of the outermost zone. The spot size limits the spatial coherence of the illuminating beam and hence the resolvable spot size of the reconstructed object. In order to produce as fine a pattern or spot size as possible, the metal foil FZP lens is formed by milling the foil using a focused ion beam (FIB) in order to produce the perforations. The resolution and accuracy of the FIB are critical to fabricating an FZP that shapes the beam properly. Some FIB instruments are capable of producing geometries that are accurate on scales ranging from 1 nm-10 nm.

    (39) The FZP efficiency can be defined as the fraction of the incident radiation that is directed as intended. For a focusing lens, that would be the fraction of the beam contained within the primary or first order spot. The diffraction order is defined as the average phase step from one transmission zone to the next in waves. Thus the zero-order is the undiffracted beam. The first order is the one where the average phase difference from one zone to the next is one wave, and so on. According to diffraction theory, a certain fraction or percentage will diffract into the different orders.

    (40) For a zone plate that is 50% opaque, efficiency cannot be greater than 50%. Since this type of diffractive optics will diffract light into orders other than the desired one, typical diffraction efficiency into the first order is 5-30%. While that results in the loss of the majority of the coherent EUV beam, it does focus part of the collimated beam and result in a diverging beam after the focus. The object 300 (i.e., the mask or mask blank) can be positioned along the optical axis so that the size of the illuminated area is close to the resolvable size of the object. Even with diffraction losses, this approach can be significantly more efficient than one that does not use optics to shape the beam. For example, it can be much more efficient than if a large, collimated beam was incident on a small object and required a masking aperture.

    (41) FZP optics offer flexibility that is difficult to achieve with more conventional optics. Namely, it is simple to add a tilt to the beam diffracted by an FZP in order to steer the beam in a particular direction. This distorts the centrosymmetric zone pattern to one that is symmetric about the projection of the tilted beam, as shown in FIG. 3. In addition, FZP optics enable simple fabrication of a compound lens where the original beam is divided into multiple beams, each of which can have a different focal length and beam angle.

    (42) The design of the FZP used for zone plate 208 is straightforward once the phase of the beam entering and exiting the FZP has been defined. The boundaries between zones are defined by locations where the optical path length of the rays connecting source and destination change by a multiple of λ/2. The design algorithm for beams that focus to, or expand from, a point source is as follows: 1. Identify the point source(s) location(s) and relative magnitude(s). Over the area of the planar (usually circular) zone plate, calculate the sum of the complex amplitude of the spherical wave(s) from the point source(s), where

    (43) U ( r 0 ) = U 0 e ikr 0 r 0 ( 4 )  is the complex amplitude of such a spherical wave, in order to calculate the complex amplitude required to produce those spherical wave(s). 2. If the beam incident on the FZP is not a plane wave, multiply the result by the complex conjugate amplitude of the input wave, in order to cancel the phase of the incident beam. 3. Select regions with phases that range over an interval [θ,θ+π) where θ∈[0,2π). Set those regions to amplitude 1 and all others to 0. 4. Adjust the design so it can be fabricated in a self-supporting fashion.

    (44) FIG. 3 shows examples of on-axis, off-axis, and compound off-axis FZPs for plane wave beams generated with the method outlined above. The plates are all defined over a 512×512 array corresponding to 13.8 μm×13.8 μm with a sampling grid of 27 nm. The point sources are all 65.8 μm distant from the FZP. The off-axis points are designed to be 2.30 μm and 4.59 μm from the center of the array along the x axis. The plates were designed for λ=13.5 nm. Using these parameters, spherical waves incident on the FZP plane were calculated and binarized as in steps 1-4 above.

    (45) The effect of these ideal FZPs can be calculated using scalar diffraction using the angular spectrum method (See J. W. Goodman, Introduction to Fourier Optics, 2nd ed., McGraw-Hill, 1996). Using that method, it is straightforward to calculate the diffraction spot at a given distance from the FZP. FIG. 3 shows calculated spots at 0.97× the focal distance and at the exact focal distance. The spots in the focal plane are quite compact.

    (46) As expected, the spot in front of the focal plane (the middle row) shows a larger spot as the beam has not yet reached the focal plane. Also, as expected, the center of mass of the spot moves in the x direction with increasing distance. The centroid of the spot in the middle row plots has not yet reached the designed position.

    (47) Binary diffractive optics are not very efficient. Roughly half the incident beam is blocked and significant portions of the transmitted beam are scattered into other orders. The angular spectrum propagation gives a first-order estimate of the diffraction efficiency of the FZP as the energy in the central spot. For the centered, off-axis, and dual off-axis examples in FIG. 3, the calculated efficiencies are 9%, 9%, and 7.5% (the total of both spots), respectively. A better estimate would require sampling the FZP with higher resolution and/or including vector diffraction effects but is not critical to engineering an APMI system since a safety factor can be built into the beam power.

    (48) Note that the design algorithm does not explicitly account for the stencil process. The shape of the stenciled holes and the fraction of open area are variables that will affect the quality of the beams U.sub.0 and U.sub.1 emerging from the zone plate 208. It is straightforward to calculate the scalar diffraction of a particular design rule for the stenciled holes and assess its performance. It may be useful to use iterative algorithms similar to those used to design phase plates that propagate the beam(s) between the plane of the zone plate (i.e., FZP 208) and one or more other planes, applying constraints on the amplitude or phase in each plane, to design the stenciled holes (see “Three-plane phase-only computer hologram generated with iterative Fresnel algorithm,” M. Makowski; M. Sypek; A. Kolodziejczyk; G. Mikula, Opt. Eng. 44(12), 125805, Dec. 29, 2005).

    (49) Several aspects of the optical system shown in FIG. 2 relate to the practicality of assembling the device. For simplicity, the beam 206 emitted by EUV source 100 is shown as perfectly collimated, turned by fold mirror R1 and split into two beams U.sub.0 and U.sub.1 (i.e., first beam 202 and second, different beam 204) by the compound FZP 208. However, as the size of the FZP is likely to be limited by strength of the foil and time/expense to machine it with a FIB, it may be useful for R.sub.1 (or another upstream mirror) to partially focus the beam 206 so that beam 206 is small when it encounters the FZP 208.

    (50) The system in FIG. 2 requires at least two mirrors (i.e., mirrors R2 and R3) to redirect the reference beam U.sub.1 (i.e., second, different beam 204). This is necessary to both clear the edge of the FZP 208 and to equalize the path length of each beam to the detector (i.e., imaging sensor 500) in order to keep them sufficiently coherent so the interference pattern has good contrast. This is challenging as the separations are fairly small. For the imaging scenario in FIG. 2 with 0.1 NA beam and angle of 0.7 mrad between the beams, R2 and R3 are separated by ˜200 μm and they sit ˜100 μm off the mask 300. Note that both beams come to focus after reflection by the mask 300 and mirror R2. This maximizes the clearance for mirror R3 so that it does not interfere with the illuminating beam U.sub.0 (i.e., first beam 202). The clearance is also increased with increasing angle between the interfering beams.

    (51) FZPs diffract some light into the first (desired) order and some light into the zero (undiffracted) order and into higher orders. Most optical systems with an FZP include an order sorting aperture near the focus of the desired (first) order to absorb light from the other orders. For a reflective dual beam system such as that shown in FIG. 2, there is no natural place for an order-sorting aperture. For the system in FIG. 2, the opening between mirrors R2 and R3 in conjunction with some absorbers can serve to eliminate most of the light in other orders. The fabrication of the optical assembly can be challenging since the clearance from the bottom end of mirrors R2 and R3 to the mask 300 is small.

    (52) One possibility is to fabricate a pyramidal assembly with the inside surface having reflective regions near the tip of the pyramid corresponding to mirrors R2 and R3 which are shown in the detailed inset A (circle) in FIG. 2. These mirrors are shown schematically with small thickness. An actual pyramidal assembly will likely have greater thickness and may require a reverse bevel so it does not hit the mask. A small clearance hole close to the mask could then be drilled with an ultrafast laser since that is a gentle way of removing the material for the hole. The inside surfaces of the assembly except for the reflecting surfaces should be EUV-absorbing, thereby absorbing most of the light from other orders.

    (53) The optical system in FIG. 2 has two reflections for the reference beam U.sub.1 (i.e., off mirrors R2 and R3) and one reflection for the illumination beam U.sub.0 (i.e., off mask 300). Thus, the handedness or parity of the two beams U′.sub.0 and U.sub.1 are not the same and, in particular, the two beams U′.sub.0 and U.sub.1 are flipped in x. This could increase the spatial coherence requirements of the beams. Other mirror systems are possible which have two reflections for both beams. Such configurations may ease the coherence requirements but have less clearance than the configuration shown in FIG. 2. Moreover, adding additional optical elements can reduce the throughput, requiring greater beam power, and may add aberrations.

    (54) Another variation of the optical system 200, shown in FIG. 2A, may also have some advantages. For example, it is desirable for the optical system to work with an EUV mask having a pellicle. A pellicle is a very thin transparent window positioned off the mask surface and parallel to the mask surface. It serves as a dust shield so dust particles which would otherwise fall onto the mask fall on the pellicle instead. Since the dust particles are not in the mask plane, they are out of focus when imaging onto the semiconductor wafer and do not spoil the lithographic pattern. In order to inspect a mask with a pellicle, a nearly collimated beam must illuminate and reflect off the mask. This beam may be designated as the propagated complex mask pattern (PCMP).

    (55) The optical system shown in FIG. 2A accomplishes this as follows. A nearly collimated coherent EUV beam 206 is split by a first FZP 210, which may be a diffraction grating, into two beams, a reference beam 214 and an illumination beam 212 for illuminating the mask 300. The reference beam 214 reflects off a mirror RR1 and the illumination beam reflects off the mask 300. The mirror RR1 and mask 300 are inclined so the beam reflects out of the initial plane defined by the beams. The beams are reflected off additional mirrors, RR2, RM2, RR3, RM3 (where the 2nd letter refers to the Reference or Main illumination beam, respectively). For convenience, these mirrors guide the beams in a plane that is parallel to the initial plane and offset normal to the initial plane until they are due to cross at 90 degrees to each other.

    (56) Just before the beams cross, they are focused by identical FZP lenses 216 and 218. This is done to minimize their cross-section so they can cross and be reflected towards the imaging sensor and to form an expanding beam to match the size of the imaging sensor. The final two mirrors, RR4 and RM4 are inclined to reflect the beams normal to the initial plane towards the imaging sensor 500. Similar to FIG. 2, these mirrors have a small reflective area in order to reduce reflections from other diffraction orders that are directed towards the imaging sensor.

    (57) With the approach shown in FIG. 2A, because each beam passes through two diffractive elements, this approach has greater diffraction losses than the system shown in FIG. 2. Each beam is also reflected by multiple mirrors which also reduces the light throughput. The optical system shown in FIG. 2 is believed to be the most optically efficient configuration with the fewest elements.

    (58) Both the optical system shown in FIG. 2 and in FIG. 2A provide essentially equal path lengths for both the illumination beam and the reference beam. This is required in order to maximize the coherence of the interfering beams and hence the contrast of the DI. The relative path lengths of the expanding illumination and reference beams also affect the interference pattern, which may affect the algorithm used to recover the mask pattern.

    (59) Another aspect of the construction shown in FIG. 2 is that the illuminating beam is shown inclined relative to mask 300 but square to the detector (i.e., imaging sensor 500). This can introduce some keystoning and also complicate the back propagation to recover the complex mask reflectivity (CMR). The detector can be tilted so it is parallel to the mask, which should eliminate the keystoning and simplify back propagation. However, that introduces path length differences across the detector (i.e., imaging sensor 500). These path length differences could impose more stringent coherence requirements on the input EUV beam than if the mask were normal to the beam. Engineering analysis is required to identify the tradeoffs resulting from the beam reflecting off the mirror at an angle and generate a design that produces the highest quality images.

    (60) The design discussed uses a 0.1 NA beam to illuminate the mask. It is possible to increase the NA and hence improve the native resolution of the system. This will require finer resolution for the FZP and may reduce the clearance for the mirror assembly. The practical limit for the beam NA will depend on the engineering details.

    (61) It will be appreciated that optical configurations other than those shown in FIG. 2 and FIG. 2A may be used to produce the interfering beams at the imaging sensor 500. For example, a system designed for transmissive objects will have some of the same constraints as one designed for reflective objects, e.g., that the two beams must be inclined by a small angle at the imaging sensor. However, that system must have the beams well separated when they pass through the transmissive object. An exemplary system would use a linear diffraction grating, either reflective or transmissive, to produce beams at angles to each other and select 2 orders, e.g., +1 and −1, while blocking all other orders. Using flat mirrors, one beam would be directed through the transmissive object and the other around the object.

    (62) The beams could then be recombined with one of two techniques. With the first technique, a second diffraction grating could be used to direct each beam by the same diffraction angle. If one beam is incident at an angle that differs by a small angle from the other beam, the diffracted beams would differ by that same small angle at the imaging sensor.

    (63) The second technique resembles the system shown in FIG. 2A. After the illumination and reference beams have been directed through the transmissive object and around it, respectively, both beams could be focused with FZPs lenses similar to 216 and 218 and directed by mirrors towards the imaging sensor. As with the optical system of FIG. 2A, the beams must be close together at the reflection point so they are directed towards the imaging sensor at a small angle to each other.

    Motion System 400

    (64) In discussing the motion system, the optical axis shall be referred to as the Z axis and the plane perpendicular to the Z axis shall be referred to as the X-Y plane. The mask inspection system translates the mask relative to a fixed illumination and imaging system with a precision motion system 400. The motion system 400 uses an X-Y table which must be fast and accurate enough to reliably move the mask 300 so that the entire surface can be imaged. DUV inspection tools typically use time delay integration (TDI) sensors. These systems continually translate the table in the fast scanning direction at a uniform speed while continually shifting the image signal across the sensor. The motion and readout systems are slaved so each pixel output of a given row of the sensor corresponds to a single pixel on the mask.

    (65) An EUV inspection system is designed to image smaller pixels at the mask than a DUV system. To inspect the same area in the same amount of time, the average pixel readout rate must increase by the ratio of the pixel areas. For an EUV system using TDI imaging sensors, the requirements to coordinate the motion and readout so that there is no smearing of the pixel data must have higher bandwidth and accuracy than for a DUV system. In addition, the focus tolerance for the mask position along the optical axis is much tighter due to reduced depth of focus for the same NA.

    (66) An EUV inspection system that does not use continuous scanning (i.e., does not use a TDI sensor) has to reposition the mask in between each exposure. Although each move is a short distance, the motion system 400 must have sufficient acceleration to move and settle the mask in a fraction of the ˜38 ms per exposure. This imposes significant requirements on the acceleration and settling time of the motion system, as discussed below. In addition, the position readout must be better than the resolution target and the readout timing must be slaved to the position. There are stages whose published specifications are close to the required acceleration and accuracy, and their applicability to the required motion profile is discussed below.

    (67) Some aspects of the motion system other than its X-Y motion may affect its suitability for inspection. For example, motion in one direction could result in displacement in another direction, i.e., motion of the X-Y stage may displace the payload in the Z direction (that is, along the optical axis). If the non-flatness is too large, it may need to be compensated by some other part of the system.

    (68) The EUV mask size, required resolution, and frame rate impose requirements on the motion system for APMI. Consider the nominal imaging specifications: an EUV mask with 132 mm×104 mm useful area and a system which images a separate 110 micron square region with each exposure with 54 nm resolution. This system requires 1.12 million exposures to cover the mask. If an inspection takes 12 hours, this requires 1 image every 0.038 s, a 26 Hz image rate. If half that time is spent performing the exposure and half that time is spent moving the mask relative to the optics, each move happens in 0.019 s. A representative timing diagram for the motion system 400 is shown in FIG. 5.

    (69) Assuming the motion profile is constant (maximum) acceleration for the first ½ of the move, followed by constant (maximum) deceleration for the second ½ of the move, the mask must accelerate at 1.21 m/s.sup.2, or 0.12 G where G is the acceleration of gravity. The maximum speed attained at the midway point is 11.6 mm/s. This can easily be achieved by existing stages from manufacturers such as Newport and Dover Motion. For example, the Newport DynamYX® DATUM® Ultra-High Performance Air Bearing Stage has 25 nm repeatability, 200 nm accuracy, accelerates at 3 G in one horizontal direction and 5 G in the other, and has a maximum speed of 1000 mm/s.

    (70) A specification that can be more difficult to meet is the requirement for the stage to settle within the resolution limit. Although there does not appear to be an off-the-shelf system with published specifications for performing the required motion and rapidly settling within a few nm, there are systems which are close, such that it is clear that a customized system meeting all the specifications can be produced (see B. Kramer, NanoPWM, January, 2016 https://www.acsmotioncontrol.com/presentations).

    Imaging Sensor 500

    (71) In order to generate a high signal-to-noise ratio (SNR) estimate of the CMR, the DI must recorded with high fidelity. As noted above, the effective angle between the interfering beams U.sub.0 and U.sub.1 varies inversely with the pixel size. Smaller pixels and a larger angle relax some of the constraints on the optical system 200.

    (72) Conversely, the requirement to record the intensity with high SNR leads to increased pixel size inasmuch as the pixel well depth typically increases with size. Consider a scenario where the average illumination is 50% of the pixel well depth. Each EUV photon, with energy 91.84 eV, produces 27 electrons when detected in silicon. For a well depth of 200,000 electrons (e), each pixel detects 3703 EUV photons, on average. Due to Poisson statistics of the detection, on average each pixel has shot noise of √{square root over (3703)}=61. This yields an average SNR of 61.

    (73) The required beam power can be estimated. Assuming a 200,000 e.sup.− well depth, 50% average illumination, a 4K×4K array, and a 7.5% efficiency of the zone plate 208, each exposure requires 7.61E13 eV or 1.22E-5 Joules. With the 26 Hz image rate and a 50% duty cycle for useful illumination (while the mask is stationary), this requires 0.63 mW of coherent EUV power.

    (74) Existing imaging detectors with suitably small pixels may not have large enough well depths to provide the desired SNR. However, techniques exist for fabricating a detector with 4.5 μm pixels with 150,000 e.sup.− well depths (see Mizobuchi, K. et al 4.5 μm Pixel Pitch 154 ke.sup.− Full Well Capacity CMOS Image Sensor. In Proceedings of the International Image Sensor Workshop, Bergen, Norway, 25-28 Jun. 2009; pp. 235-238). Therefore, a custom detector can be constructed for an APMI system which would provide the desired SNR.

    (75) According to the nominal scenario described above, with 26 4K×4K exposures per second and each pixel having 16 bit readout, the imaging sensor must deliver 872 GB/s. Assuming equal times to expose and read out the sensor, the data rate is 1.745 TB/s. Both the array size and readout rate are well within industry capabilities as exemplified by 4K HDTV imaging sensors, e.g., the Sony sensor in the HDC-4800 camera with ˜8 Mpixels can be read out at 480 frames per second (fps) which is equivalent to 4 Gpix/s.

    Algorithm 600

    (76) There are a variety of means to form an image of the mask which is efficiently illuminated by a coherent EUV beam. For example, imaging optics, e.g., a FZP or reflective imaging optics, could be used to form a conventional image of the mask 300 onto imaging sensor 500. This approach is straightforward, but does not readily lend itself to easily imaging phase defects or scanning the entire mask without experiencing focus shifts. A preferred embodiment of the present invention uses the following algorithm.

    (77) The imaging sensor or detector 500 records the intensity image of the interference pattern of the two beams U.sub.0 and U.sub.1, which is denoted as the diffraction image (DI). To understand how the image of the mask reflectivity is recovered, consider the amplitude (Eq. 5) and intensity (Eq. 6) of coherent beams incident on the detector, given by the following equations:
    U.sub.D(r)=U′.sub.0D(r)+U.sub.1D(r)  (5)
    and
    I.sub.D(r)=I.sub.0D′(r)+I.sub.1D(r)+U′.sub.0D(r)U.sub.1D*(r)+U.sub.0D*′(r)U.sub.1D(r)  (6)
    where U.sub.D is the complex amplitude at the detector and I represents the intensity due to the illuminating beam (0), reference beam (1), after propagation, with the apostrophe ′ indicating that reflection by the mask has modified the beam.

    (78) Taking the Fourier transform of Eq. 6 gives the following:
    custom character{I.sub.D(r)}=custom character{I.sub.0D′(r)}+custom character{I.sub.1D(r)}+custom character{U.sub.0D′(r)U.sub.1D*(r)}+custom character{U.sub.0D*′(r)U.sub.1D(r)}  (7)
    The first two terms are centered around zero frequency, as they are the Fourier transform of the autocorrelation of the amplitudes of the two beams. The next two terms are Fourier transforms of the cross-terms and can be separately recovered if the two beams are properly designed.

    (79) To see this, consider the ideal case where the mask is a perfect mirror (P.sub.M=1) and the two beams are nearly plane waves at the detector, so they are approximated as follows:
    U.sub.0D=A.sub.0e.sup.−ik.sup.0.sup.x
    U.sub.1D=A.sub.1e.sup.−ik.sup.0.sup.x  (8)
    where k.sub.0=2π cos θ/λ is the angular spatial frequency for a beam inclined (symmetrically in x) by θ from the normal to the detector and without loss of generality A.sub.0 and A.sub.1 are real.

    (80) The first two terms of Eq. 7 reduce to the Fourier transform of a constant, which is a delta function at zero frequency. To evaluate the next two terms, use the relation
    custom character{e.sup.−ik.sup.0.sup..Math.r}=δ(k+k.sub.0)  (9)
    and the convolution theorem, which states that the Fourier transform of a product of functions is the convolution of the Fourier transform of the functions,
    custom character{fg}=custom character{f}.Math.custom character{g}  (10)

    (81) For the beams in Eq. 8, Eq. 7 reduces to
    custom character{I.sub.D(r)}=A.sub.0.sup.2δ(0)+A.sub.1.sup.2δ(0)+A.sub.0A.sub.1*δ(k+2k.sub.0custom character)+A.sub.0*A.sub.1δ(k−2k.sub.0custom character)  (11)
    Thus, the contribution from the interference terms has been shifted to ±2k.sub.0custom character. This result is equivalent to the fact that the plane waves interfere to produce a perfect sinusoidal fringe pattern with
    I.sub.D(r)=A.sub.0+A.sub.1+2A.sub.0A.sub.1 cos(2k.sub.0x)  (12)

    (82) For the case where the mask has a pattern, Eq. 7 becomes
    custom character{I.sub.D(r)}=A.sub.0.sup.2δ(0)+A.sub.1.sup.2δ(0)+A.sub.0A.sub.1custom character(k+2k.sub.0custom character)+A.sub.0A.sub.1custom character*(k−2k.sub.0custom character)  (13)
    where custom character represents the propagated complex mask pattern (PCMP) in the detector plane and custom character is its Fourier transform. In this case, the Fourier transform of the PCMP or its conjugate are shifted to ±2k.sub.0custom character.

    (83) The method of recovering the mask pattern is straightforward (see M. Liebling, T. Blu, and M. Unser, “Complex-wave retrieval from a single off-axis hologram,” J. Opt. Soc. Am. A 21, 367-377, 2004). The Fourier transform of the detector intensity image is calculated. The regions of the transform around zero frequency and the positive frequency 2k.sub.0custom character are zeroed (masked off). The remaining region around the negative frequency −2k.sub.0custom character is inverse transformed to recover the propagated mask pattern. This recovery algorithm is termed Fourier-masking (FM) which produces an estimate of the PCMP. The PCMP estimate is then back-propagated to the mask plane which produces an estimate of the complex mask pattern P.sub.M which is denoted as the complex mask reflectivity (CMR).

    (84) Once a given area of the mask has been imaged, the mask can be translated (i.e., moved) parallel to its surface to image a different area of the mask. A series of DIs are recorded and processed to produce the corresponding CMRs. The CMRs can be stitched together to estimate the complex reflectivity of the entire mask pattern.

    (85) For an optical system like that of FIG. 2, the two beams are expanding spherical waves with finite extent, not the plane waves of Eq. 8. This changes the result in Eq. 13 in several ways. First, the first two terms are no longer delta functions. Instead, they are the Fourier transform of the autocorrelation of the beams. Due to the finite extent of the beams, the Fourier transform of the propagated mask pattern is now convolved with a function having some width, so the propagated mask pattern recovered by the FM algorithm is slightly degraded. However, if the beams roll off smoothly, the distortion of the CMR is only on the large scale of the beam roll off.

    (86) A more significant change is that the interfering spherical waves deviate from the ideal plane waves and produce hyperbolic fringes. With increasing NA, the fringes (from a flat mask) increasingly deviate from linearity. The requirement to have high resolution at the mask drives the NA to be as large as possible. As is illustrated below, the effect of non-linear fringes can significantly degrade the resolution of the PCMP and the back-propagated CMR. However, there is a technique that significantly mitigates this effect.

    (87) The recovery algorithm performs two major calculations. The first is the FM estimate of the PCMP. The second is the back-propagation of this result to the mask plane to estimate the CMR. As these estimates of the mask reflectivity and the wafer exposure are numerical calculations, they can accommodate certain types of instrument variation more easily than conventional aerial imaging. For example, parameters such as the longitudinal (focus) position of the mask become software parameters that can accommodate variations as the mask is translated (i.e., moved) to image the entire surface.

    (88) The estimate of the CMR is the main result of APMI. It can easily be compared with the designed mask pattern (amplitude), inspected for particles and other defects, and the phase can indicate how well the mask matches the ideal flat phase which mask blanks are meant to produce. However, an additional calculation must be performed to produce a result like that of aerial imagery. The complex reflectivity must be used with a model of the exposure tool to calculate the intensity at the wafer. This process is not the same as directly measuring the aerial image but it can be more powerful. To the extent that there are parameters of the exposure tool which can be varied, or if aberrations of a particular exposure tool are known, this approach can predict the aerial image with a particular set of exposure parameters or a particular exposure tool.

    (89) According to the discussion above, this is the basic algorithm for recovering the CMR: 1. Take the fast Fourier transform (FFT) of the DI. 2. Zero out (mask off) the regions of the transform around zero frequency and the positive frequency 2k.sub.0custom character. 3. Take the inverse FFT to recover the PCMP. 4. Back propagate the PCMP to estimate the CMR.

    (90) As mentioned above, interfering spherical waves produce hyperbolic fringes, not straight fringes as are implicit in Eqs. 10-12. The nature of the problem is demonstrated by performing a numerical simulation which is shown in FIG. 6. The off-axis lobes of the Fourier transform of the hyperbolic fringe pattern are not delta functions. Instead, they are compact functions with some shape. Referring to Eq. 7, the PCMP is convolved with this compact function. It is desirable to straighten the fringes to make the off-axis lobes more point-like. Doing so increases the fringe visibility/SNR and simplifies the process of recovering the PCMP.

    (91) As is seen in FIG. 6, adding an (x.sup.4+y.sup.4) term (spherical aberration) to the interfering spherical waves significantly straightened the fringes so the off-axis lobes of the Fourier transform are closer to an ideal delta function. In generating the fringes in FIG. 6, each aberrated spherical wave was formulated as follows,
    U(r.sub.0)=U.sub.0/r.sub.0 exp(ik√{square root over (r.sub.0.sup.2+(fac)(x.sub.0.sup.4+y.sub.0.sup.4))})  (14)

    (92) For the optical system of FIG. 2, this correction can be implemented as follows. The FZP 208 producing the dual beams U.sub.0 and U.sub.1 can be generated using an aberrated spherical wave. The spherical wave of Eq. 4 is replaced by an aberrated wave formulated as in Eq. 14. The corresponding fringe pattern is then calculated in the detector plane and the value of the aberrations optimized to produce point-like off-axis lobes in the Fourier transform of the fringe pattern.

    (93) It is straightforward to generalize Eq. 14 to use additional aberrations. Because the fringes are not radially symmetric, using asymmetric aberrations like astigmatism and coma should result in straighter fringes. This would yield more compact off-axis lobes than the example shown in FIG. 6, and hence a more accurate estimation of the PCMP and CMR.

    Computational and Data System 700

    (94) As noted above, the nominal system produces data at a peak rate of 0.84 Gpix/s. Assuming 16 bit pixels, this is 13.4 Gbit/s. This is well within the data transfer and data storage rates of modern computer buses and storage systems, if the system is designed to store all data in real-time.

    (95) It may be preferable to compute the CMR in real-time. This can be accomplished using graphics processing units (GPUs) as special purpose processors. There are many GPUs available which can perform TFlops/s, for example the NVIDIA TITAN RTX which delivers 130 TFlops (see https://www.sensorsmag.com/embedded/gpu-delivers-130-teraflops-deep-learning). GPUs have been used to perform the FFTs used in these algorithms (see H. Bauke, C. Keitel, “Accelerating the Fourier split operator method via graphics processing units,” Computer Physics Communications Volume 182, Issue 12, Pages 2454-2463, 2011) and can be programmed to implement the FM algorithm and back-propagation. The engineering of a real-time system is complex, especially when one considers that some parameters of the processing will be adapted to the data. However, data and computation systems with the necessary bandwidth and computing power are known to those skilled in the art.

    Simulation

    (96) A numerical simulation of the APMI imaging system of the present invention was performed to evaluate the performance of the system and in order to engineer the data system and algorithms. The simulation assumes that a coherent beam from EUV source 100 passes through a compound FZP 208 which generates two perfect spherical waves, U.sub.0 and U.sub.1. One beam (U.sub.0) is reflected by the mask CMR P.sub.M and propagated to the detector plane. The other beam (U.sub.1) is returned to the detector plane via mirrors R.sub.2 and R.sub.3. The two beams U.sub.0 and U.sub.1 received at the detector are added and the intensity image is calculated. The FM algorithm described above is applied to the intensity to estimate the PCMP. The resulting PCMP is numerically propagated back to the mask plane to estimate the CMR. The following steps were performed: 1. Generate a sample image of the USAF resolution target. 2. Calculate the amplitude and phase of two spherical waves U.sub.0 and U.sub.1 at the detector, whose center points are offset by a small distance resulting in a small effective angle of incidence at the imaging sensor. 3. Propagate the sample image P.sub.M from the mask plane to the detector. 4. Calculate the intensity at the detector, |custom characterU.sub.0+U.sub.1|.sup.2, and smooth the edges by multiplying with a super Gaussian factor. 5. Calculate the FFT, mask off the appropriate off-axis pattern, i.e. set all portions of the FFT to zero except the appropriate off-axis pattern, then calculate the inverse FFT to estimate the PCMP. 6. Numerically back propagate the PCMP estimate to estimate the CMR. Here is a description of the steps in more detail.

    (97) Step 1. An ideal binary image of the USAF resolution target was spatially filtered in order to simulate finite resolution of the fabrication process. FIG. 7 shows the smoothed target, a 2048×2048 array with 54 nm pixels.

    (98) Step 2. In order to separate the signals in frequency space, the two spherical wave origins were offset in x with a separation that produces a fringe pattern with a 3 pixel period at the detector. The detector pixels were assumed 6 μm. With NA=0.1 illumination, a propagation distance of 41 mm appropriately spread the propagated image across the detector and set the separation of the sources to 31 μm. U.sub.0 and U.sub.1 were calculated as spherical waves, i.e., there was no paraxial approximation. Considering the interference of the spherical waves, FIG. 8 shows the Fourier transform of the intensity pattern. FIGS. 8a and 8b show log.sub.10 of the transform for pure spherical waves. FIG. 8c shows log.sub.10 of the transform for waves with a (x.sup.4+y.sup.4) term that straightens the fringes.

    (99) Note that for simplicity, the two beams are displaced along the X axis or angled about the Y axis. It is preferable to offset the beams at 45° to the imaging sensor axes since that maximizes the available area in frequency space for the off-axis lobes, thereby increasing the resolution. For a system such as that shown in FIG. 2, this can easily be accomplished by rotating the detector by 45° relative to the fringes or displacement vector.

    (100) Steps 3-4. Using available code, the USAF pattern was propagated using a split-step Fresnel propagation. The Talanov lens transformation was used to take the spherical wavefront illumination into account (see V. I. Talanov: “Focusing of light in cubic media,” JETP Letters 11, 799, 1970 [ZhETP Pis. Red. 11, 303 20 (March 1969)]; Fleck, J. A., Morris, J. R. & Feit, M. D., “Time-dependent propagation of high energy laser beams through the atmosphere,” Appl. Phys. Volume 10, pp 129-160, 1976). The pixel grid at the sample image and propagation distance are chosen to match the implicit magnification by this method. The PCMP and the fringe pattern resulting from |custom characterU.sub.0+U.sub.1|.sup.2 is shown in FIG. 9. FIG. 9b represents the output of the imaging detector.

    (101) Steps 5-6. The FFT of the interference pattern, e.g., FIG. 9b, is calculated. The sidelobe used to recover the PCMP is masked off. FIG. 10 shows the outputs at various steps in the algorithm. In the example shown, a circular mask with the diameter ⅓ the width of frequency space was used. The masked off FFT is inverse transformed to estimate the PCMP. This complex function is then back-propagated to estimate the CMR. As expected, adding aberrations to straighten the fringes improves the resolution of the output, especially in the direction perpendicular to the fringes, as is seen in the improvement from FIG. 10e to FIG. 10f. FIG. 10f is not quite as sharp as FIG. 7b. This is due to the spatial frequency filtering of FIG. 10b. Applying equivalent filtering to FIG. 7b produces an image the same as FIG. 10f.

    (102) The phase of the recovered image is shown in FIG. 11. The input image had the phase equal to zero. The recovered image has nearly flat phase over each bar, but there is some shape to the phase, presumably due to the deviation of the fringes from perfectly straight ones as seen in FIG. 6. The recovered image had to be corrected for linear phase ramps due to the angle between the beams and resulting position of the off-axis lobes.

    (103) This simulation is representative of the sort of imaging capable with the apparatus and method described herein. However, some discrepancies from the actual imaging scenario shown in FIG. 2 exist, which could point to differences between the simulation and results of a real system. They are listed in decreasing order of importance: 1. The simulation uses the Fresnel approximation for propagation from the mask to the detector. Within the limits of scalar diffraction, the spherical component of the wavefront is accurately modeled but the propagation relies on an approximation. Depending on the NA of the beams, a more accurate propagator may be required to recover the CMR with highest accuracy. 2. The simulation assumes the illumination and reference beams are ideal spherical waves emitted from a point source. The system outlined in FIG. 2 differs in 2 ways. First, the beam is generated by a finite diameter FZP focusing a plane wave to a diffraction limited spot. This will limit the resolution of the system based on the NA of the beam as the resulting illumination is equivalent to a superposition of point sources over a diffraction limited spot whose size is set by the wavelength and NA. Second, the converging U.sub.0 beam reflects off the mask before passing through focus. This likely introduces non-uniform illumination of the mask, which can likely be calibrated. 3. The simulation assumes the reference beam point source is symmetrically offset from the illumination beam point source about the axis defined by the center of the object (i.e., the mask) and the imaging sensor. Their path lengths to the imaging sensor are nominally exactly equal. A deterministic small offset can be added to simulate non-ideal assembly. 4. Propagation of the beams is always in the same (forward) direction and the mask and imaging sensor are in parallel planes. The mask is assumed to be an infinitely thin pattern that multiplies the complex amplitude of the incident beam to modulate its magnitude and phase. Thus the inclination of the illumination beam incident on the mask plane is not simulated, so issues of keystoning or propagation back to an inclined plane are not considered. It also assumes the beams are perfectly coherent.

    (104) A more complete simulation can be constructed which includes all the non-ideal features of a real device, but this simulation models the features of the proposed invention which directly bear on its utility. Aside from being a limited proof-of-concept, the imaging simulation is useful in engineering an APMI system. By varying certain parameters, the potential resolution and accuracy of a system can be probed, allowing some engineering to be done by running simulations. In addition, the effect of non-idealities can be simulated in order to validate the sensitivity of the approach. For example, a random phase due to non-flatness of the mask blank, path length mismatch between the illumination and reference beams, focus drift of the system, detector noise and detector non-uniformity can all easily be simulated to probe their effect on the recovered CMR.

    (105) Exercising the simulation has quantified some of the sensitivity of the system. For example, it was found that a path length mismatch of 1 micron between the reference beam U.sub.1 and the illumination beam U.sub.0 significantly limited resolution for the 0.1 NA case described above. In addition, the path length to the mask for the back propagation must be known to better than 100 nm in order to produce clean images at the highest resolution. Since that value is not likely to be known to that accuracy from the mechanical construction, it must be found after assembly by optimizing the image of a known object.

    Some Unique Features of the Present Invention

    (106) The approach of the present invention differs from other approaches in several ways. In contrast with conventional aerial imaging, the approach of the present invention does not require that the mask be maintained at an absolute Z (focus) position with nm precision. If the Z position of the mask varies as the surface is scanned, it is possible to compensate for this by varying the Z distance used when back propagating the PCMP. The Z distance can be optimized by various measures, e.g., the sharpness of the recovered image or its match to the known mask pattern, and the optimized value used for the recovery. The Z position will vary smoothly as the mask is scanned. This smooth variation can be fit in software, and will likely repeat from scan to scan, so the software burden for tracking it will be minimal.

    (107) As this is a coherent imaging method, it provides the modulus and phase of the mask reflectivity. In contrast to CDI, this method requires no iteration—the FM algorithm and subsequent back propagation are direct methods. This enables reliable engineering of a real-time imaging system or a post-processing system, since the computational burden to produce an image is fixed.

    (108) Another distinctive feature of the present invention is that this imaging approach does not require curved reflective EUV optics. It requires several flat mirrors, but these are much easier and less costly to fabricate than the reflective imaging optics required for aerial imaging. However, the required coherent source of the present invention is a unique component.

    Modifications

    (109) The described embodiments of the invention are intended to be merely exemplary and numerous variations and modifications will be apparent to those skilled in the art. All such variations and modifications are intended to be within the scope of the present invention.

    (110) Among other things, it will be appreciated by those skilled in the art that the invention described herein is not limited in application to semiconductor masks or mask blanks but may be applied to imaging a variety of objects in reflection or transmission. It will also be appreciated by those skilled in the art that the invention is not limited to quasi-planar objects, but may be applied to objects where the source light interacts over some significant depth of the object. The back propagation can be done to varying distances, thereby enabling reconstruction of the object as a function of depth. These and other variations are considered to be within the scope of the present invention.