Predicting and measuring melanopic dose

11287321 · 2022-03-29

    Inventors

    Cpc classification

    International classification

    Abstract

    Melanopic dose rate and dose are calculated in a virtual environment. A computer generated model of an actual or planned building is used as the virtual environment. Indirect and direct spherical irradiances are calculated using convex polyhedra throughout the virtual environment, and each is multiplied by a melanopic conversion factor. The two are added, then adjusted for a human's angular responsivity and age. Building design features or lighting devices may be adjusted to provide a required melanopic dose rate. A camera is used to capture a panoramic image, which is calibrated to tristimulus values, and used with the spectral power distribution of the light sources to derive the melanopic dose rate.

    Claims

    1. A method for calculating melanopic dose rate for a direction of gaze at a position in a virtual environment comprising the steps of: a) calculating, by a processor, a direct spherical irradiance at the position from one or more light sources, as a function of angle; b) multiplying, by the processor, the direct spherical irradiance by a melanopic conversion factor that depends on a spectral power distribution of the direct spherical irradiance; c) calculating, by the processor, an indirect spherical irradiance at the position from multiple light source patches, as a function of angle; d) multiplying, by the processor, the indirect spherical irradiance by a melanopic conversion factor that depends on a spectral power distribution of the indirect spherical irradiance; e) adding, by the processor, the products from steps (b) and (d) to form a total spherical melanopic irradiance defined as a function of angle; and f) multiplying, by the processor, the total spherical melanopic irradiance by an angular responsivity that depends on the direction of gaze, to result in the melanopic dose rate.

    2. The method of claim 1, comprising calculating, by the processor, a melanopic dose by multiplying the melanopic dose rate by a dwell time of a virtual observer.

    3. The method of claim 1 wherein: the position is a position of a virtual meter in the virtual environment; the one or more light sources include virtual representations of direct solar and diffuse daylight; and step (c) comprises projecting each light source patch onto one or more subdivided surfaces of a convex polyhedron that encloses the position.

    4. The method of claim 3, wherein the convex polyhedron is a dual cubic tetrahedron.

    5. The method of claim 1, wherein the virtual environment is a model of a building and the method comprises adjusting, by a user, a design element in the model to modify the melanopic dose rate.

    6. The method of claim 5, comprising incorporating the adjusted design element in the building.

    7. The method of claim 1, wherein the virtual environment is a model of a building and the method comprises verifying that the melanopic dose rate is within a predetermined range.

    8. The method of claim 1, comprising: modeling, by a user, an interior of a building to result in the virtual environment; providing, by the user, the virtual environment to the processor; and obtaining, by the user via an interface connected to the processor, the melanopic dose rate.

    9. A system for calculating melanopic dose rate for a direction of gaze at a position in a virtual environment comprising: a processor; a user interface connected to the processor; and a computer readable memory storing computer readable instructions, which, when executed by the processor cause the processor to: a) calculate direct spherical irradiance at the position from one or more light sources, as a function of angle; b) multiply the direct spherical irradiance by a melanopic conversion factor that depends on a spectral power distribution of the direct spherical irradiance; c) calculate an indirect spherical irradiance at the position from multiple light source patches, as a function of angle; d) multiply the indirect spherical irradiance by a melanopic conversion factor that depends on a spectral power distribution of the indirect spherical irradiance; e) add the products from steps (b) and (d) to form a total spherical melanopic irradiance defined as a function of angle; f) multiply, the total spherical melanopic irradiance by an angular responsivity that depends on the direction of gaze, to result in the melanopic dose rate; and g) output, via the user interface, the melanopic dose rate.

    10. The system of claim 9, wherein the virtual environment is a model of a building, the system comprising a controller connected to the processor and configured to control a luminaire or a fenestration device in the building to provide a required melanopic dose rate in the building.

    11. A method to measure melanopic irradiance in a physical environment, comprising: rotating a camera about its axis within the physical environment to generate a panoramic image; calibrating the panoramic image to correspond to CIE XYZ tristimulus values of the physical environment at a position of the camera; and using spectral power distribution values of one or more light sources in the physical environment to calculate a melanopic irradiance at the position of the camera.

    12. The method of claim 11 wherein the one or more light sources include daylight, the method further comprising: including changes in the daylight due to changing sky conditions and time of day; and including effects of fenestration devices in the calculation.

    13. A system to measure melanopic irradiance in a physical environment, comprising: a calibrated panoramic digital camera; a device to rotate the camera to generate a panoramic image; a computer configured to calibrate the panoramic image to CIE XYZ tristimulus values at a position of the camera in the physical environment; a database with spectral power distribution values of light sources present in the physical environment; and a computer configured to calculate the melanopic irradiance from the CIE XYZ tristimulus values and the SPD values.

    14. The system of claim 13 comprising a database storing daylight data.

    15. The system of claim 13 comprising a database storing data of fenestration devices present in the physical environment.

    Description

    BRIEF DESCRIPTION OF DRAWINGS

    (1) FIG. 1 shows the CIE photopic and melanopic action spectra.

    (2) FIG. 2 shows the change in human corneal spectral transmittance with age.

    (3) FIG. 3 shows an empty rectangular room with one example surface subdivided into patches and one example light source.

    (4) FIG. 4 shows the spherical irradiance received at a point from a single patch.

    (5) FIG. 5 shows a dual cubic tetrahedron.

    (6) FIG. 6 shows the subdivision of a cubic tetrahedron face into an array of elements.

    (7) FIG. 7 shows the orientation of a dual cubic tetrahedron with respect to the u-v-n Cartesian coordinate system.

    (8) FIG. 8 shows a patch being projected onto the face of a cubic tetrahedron.

    (9) FIG. 9 shows a plot of the Guth position index.

    (10) FIG. 10 shows the discomfort glare angular responsivity plot.

    (11) FIG. 11 shows a flowchart presenting the calculation of spherical irradiance for a specified observer position.

    (12) FIG. 12 shows a flowchart presenting the calculation of melanopic irradiance from spherical irradiance for a specified observer gaze direction.

    (13) FIG. 13 shows a digital camera being rotated about a vertical axis to obtain a panoramic image.

    (14) FIG. 14 is a schematic diagram of an exemplary system for the measurement or prediction of melanopic dose and dose rate.

    (15) FIG. 15 shows three panoramic images with electric lighting, daylighting, and a combination of the two.

    DETAILED DESCRIPTION

    (16) In designing an illuminance meter for virtual environments, it is assumed that the meter has a specified position and orientation within the three-dimensional space. It may be similarly assumed that a melanopic irradiance meter has a specified position for the observer, but there is no reason to assume a specified orientation. For example, an office worker or student sitting at a desk for an hour or so will most likely direct their gaze in a variety of different directions in response to visual and auditory stimuli. In order to predict a physically plausible melanopic dose over this time period, it is reasonable to consider a stochastic behavioral model for head movement and gaze direction of the observer.

    (17) Spherical Irradiance

    (18) While the melanopic irradiance for a given orientation taking into account the observer's angular responsivity to said irradiance could be predicted, this becomes computationally expensive for multiple gaze directions. A more efficient approach is to measure the spherical irradiance distribution at the position, and then derive melanopic irradiances for a multiplicity of arbitrary gaze directions from this distribution.

    (19) Referring to FIG. 3, this measurement process begins by choosing at an arbitrary meter position 315 in the virtual space 300. If the position (as measured from a light source 305) is at a distance r that is approximately two or more times the diagonal width of the emission area of the light source 305, the light source can be represented as a point source of irradiation p at the center of the emission area, and the direct irradiance E.sub.p,o of the position 315 calculated using the inverse square law:
    E.sub.p,o=I.sub.p,Ω/r.sup.2  (1)

    (20) where I.sub.p,Ω is the radiant intensity of the light source 305 in the direction Ω (expressed in polar coordinates) of the meter position 315.

    (21) If the distance r is less than approximately twice the diagonal width of the emission area of the light source 305, the emission area may be subdivided into a one- or two-dimensional array of smaller subareas such that the diagonal width of each subarea satisfies the distance criterion, and where the total radiant flux Φ emitted by the light source is evenly divided between them. Each subarea is represented as a point source located at the center of the subarea.

    (22) The inverse square law calculation is repeated for each light source or subdivision thereof 305 in the virtual environment that is directly visible from the position 315 to determine the total direct spherical irradiance E.sub.o at the meter position 315. This approach is adapted herein such that the spherical irradiances E.sub.p,o and the corresponding directions Ω for all point sources p are saved in a list assigned to meter position 315. This list represents the direct spherical irradiance as a function of angle Ω at position 315, where each list element is comprised of: a) Direct spherical irradiance; and b) Polar coordinates δ (horizontal angle) and (vertical angle of light source 305 as seen from position 315.

    (23) The same calculation may be repeated for other arbitrary positions of a virtual meter in the virtual environment.

    (24) The indirect spherical irradiance at the position 315 is a result of light being reflected from surfaces in the virtual environment that are not themselves light-generating sources, such as the floor, ceiling, and walls. Each surface must be subdivided into an array of small patches (e.g., 310) and their direct irradiance from the light sources and indirect irradiance from light reflected from other patches calculated in order to determine their radiant exitance value M.sub.e. This can be done using, for example, the radiosity method described in Ashdown, I. 1994. Radiosity: A Programmer's Perspective. New York, N.Y.: John Wiley & Sons.

    (25) To calculate the indirect spherical irradiance E.sub.j from each surface patch 310 visible to point 315, and referring to FIG. 4, it is first noted that the spherical irradiance E.sub.j received at point 405 from a patch j (410) with surface area 4 and radiant exitance M.sub.e that is at a distance r.sub.j and at an angle φ.sub.j to the surface normal N is:
    E.sub.j=1/π∫.sub.ΩM.sub.e  (2)
    where the differential solid angle dω is defined as:
    dω=cos(φ.sub.j)dA.sub.j/r.sub.j.sup.2  (3)

    (26) The integration can be performed by numerical quadrature using a convex polyhedron. In this example a dual cubic tetrahedron, is used. This consists of two back-to-back cubic tetrahedrons measuring two units on each side that are centered on the point P (FIG. 5). Each face of the dual cubic tetrahedron is divided into an array of n=m×m/2 cells with area ΔA.sub.m, as shown in FIG. 6. (The cells along the adjoining edges of the two tetrahedrons have area ΔA.sub.m/2.) Thus, for a given patch j, and a given face of the convex polyhedron from which the patch j is visible:
    E.sub.j≈1/πΣ.sub.i=1.sup.nM.sub.e cos(φ.sub.j)ΔA.sub.m/r.sub.j.sup.2  (4)

    (27) FIG. 7 shows an orthographic projection of the upper cubic tetrahedron aligned on a horizontal plane with a u-v-n Cartesian coordinate system centered on point P and oriented such that each axis intersects a face of the cubic tetrahedron at a distance of unity from point P.

    (28) Taking the cubic tetrahedron face intersected by the n-axis as an example, we have for each cell i with variable coordinates u.sub.i and v.sub.i:

    (29) r i = u i 2 + v i 2 + 1 ( 5 )
    and:
    cos(φ.sub.i)=1/r.sub.i  (6)
    Thus, for a given patch j, its contribution to a given cell i in a face of the cubic tetrahedron from which the patch j is visible is:
    ΔE.sub.i≈M.sub.eΔA.sub.m/π(u.sub.i.sup.2+v.sub.i.sup.2+1).sup.3/2  (7)
    and similarly for all six faces of the cubic tetrahedron.

    (30) This can be confirmed by assuming a uniform sphere with unit radiant exitance M.sub.e surrounding the light buffer (i.e. the dual cubic tetrahedron) and centered on point P. A diffusely-emitting surface with radiant exitance M has radiance M/π in any direction, which gives:
    ΔE.sub.i≈ΔA.sub.m/(u.sub.i.sup.2+v.sub.i.sup.2+1).sup.3/2  (8)

    (31) Summing ΔE.sub.i over all n cells (in the limit ΔA.sub.m.fwdarw.0) for each side yields 4π, which is the area of a unit sphere.

    (32) Referring to FIG. 7, the cubic tetrahedron faces are oriented with respect to the u-v-n Cartesian coordinate system such that:

    (33) A = { - 1 / 6 , 1 / 2 , 0 } B = { - 1 / 6 , - 1 / 2 , 0 } C = { 2 / 3 , 0 , 0 } D = { 0 , 0 , 0 } u CT = { 1 / 6 , 1 / 2 , ± 1 / 3 } v CT = { 1 / 6 , - 1 / 2 , ± 1 / 3 } n CT = { - 2 / 3 , 0 , ± 1 / 3 } apex = { 0 , 0 , ± 1 / 3 }
    where the ± symbols are positive for the upper cubic tetrahedron, and negative for the lower one.

    (34) FIG. 8 shows a triangular patch 805 being projected onto a face of a cubic tetrahedron 810, where the projection covers the set of face cells 815, wherein the projection process is disclosed in, for example, Chapter 5 of Ashdown, I. 1994. Radiosity: A Programmer's Perspective. New York, N.Y.: John Wiley & Sons.

    (35) In an adaptation of the above, the projection of the surface patches 310 onto the six faces of the dual cubic tetrahedron are saved as six triangular arrays with irradiance values for each face cell 815 in accordance with Equation 7. Note that each face cell has a specific direction expressed in polar coordinates from the center P of the dual cubic tetrahedron. Similar to the direct spherical irradiance list, each face cell element is comprised of: a) Indirect spherical irradiance incident on the face cell; and b) Polar coordinates δ (horizontal angle) and (vertical angle of the face cell center as seen from the center of the dual cubic tetrahedron.

    (36) FIG. 11 shows a flowchart presenting the calculation of spherical irradiance for a specified observer position. In Step 1100, a radiosity solution is calculated for the virtual environment in order to determine the surface patch exitances. Assuming that the color of each surface is specified as an RGB triplet or (equivalently) as a CIE xyY chromaticity, red, green, and blue radiosity values are calculated in parallel. The light source SPD is assumed to be CIE Standard Illuminant D65 (6500K daylight), as specified in Colorimetry, Fourth Edition, CIE 015:2018).

    (37) In Step 1105, an observer position P within the virtual environment is specified.

    (38) In Step 1110, a light source is chosen from the set of light sources. (If the light source has been subdivided due to its proximity to the observer position, a light source segment is chosen.)

    (39) In Step 1115, the direct spherical irradiance at observer position P is calculated in accordance with Equation 1, along with the polar angle of the light source with respect to the observer position.

    (40) In Step 1120, a decision is made as to whether there are more light sources or light source segments to choose. If true, control passes to Step 1110 to choose the next light source or segment. If false, control passes to Step 1125.

    (41) In Step 1125, one of the face cells from the six cubic tetrahedron faces is chosen.

    (42) In Step 1130, the indirect spherical irradiance at observer position P is calculated in accordance with Equation 7, along with the polar angle of the face cell with respect to the cubic tetrahedron center (i.e., the observer position).

    (43) In Step 1135, a decision is made as to whether there are more face cells to choose. If true, control passes to Step 1125 to choose the next face cell. If false, control passes to Step 1140.

    (44) In Step 1140, a light source is chosen from the set of light sources. (If the light source has been subdivided due to its proximity to the observer position, a light source segment is chosen.)

    (45) In Step 1145, the face cell whose polar angle is closest to the light source polar angle is determined.

    (46) In Step 1150, the direct spherical irradiance due to the light source or light source segment is added to the face cell's indirect spherical irradiance.

    (47) In Step 1155, a decision is made as to whether there are more light sources or light source segments to choose. If true, control passes to Step 1140 to choose the next light source or segment. If false, the algorithm finishes.

    (48) It should be noted that upon completion, the six faces with their square face cells are conceptually equivalent to triangular digital images of the virtual environment as seen from the observer position. It is therefore possible to capture an image of a physical environment with a calibrated 360-degree spherical camera such as the Ladybug5+ manufactured by FLIR Systems, Arlington, Va., and geometrically map the image pixels to the face cells of a cubic tetrahedron.

    (49) Angular Responsivity

    (50) The angular responsivity of an observer with respect to gaze direction may be represented as an analytic function or tabulation of values at discrete angles. The simplest representation is the field of view limits presented in CIE System for Metrology of Optical Radiation for ipRGC-Influenced Responses to Light, CIE S 026/E:2018, but it is more reasonable to assume that the angular responsivity of a typical observer will be similar to that of an observer's responsivity to discomfort glare. Discomfort Glare in Interior Lighting, CIE 117-1995, for example, presents a tabulation of “Guth position indices” relative to the gaze direction (Guth, S. K. 1951. “Brightness Relationships for Comfortable Seeing,” Journal of the Optical Society of America 41(4):235-244), while Levin, R. E. 1975. “Position Index in VCP Calculations,” Journal of the Illuminating Engineering Society 4(2):99-105, presents an analytic function that was empirically derived from the tabulated values. A nomogram of the Guth position indices is shown in FIG. 9, while a hemispherical plot of the (inverse) angular responsivity to discomfort glare from Jakubiec. J. A. and C. F. Reinhart. 201. “The ‘Adaptive Zone’—A Concept for Assessing Discomfort Glare throughout Daylight Space,” Lighting Research and Technology 44(2)149-170, is shown in FIG. 10.

    (51) The gaze direction of an observer can be described in terms of three orthonormal vectors u, v, and n, where n is the gaze direction, v is the “view-up” direction for the observer, and u is the direction of the right-hand rule for vectors v and n (e.g., Ashdown, I. 1994. Radiosity: A Programmer's Perspective. New York, N.Y.: John Wiley & Sons). Given the direction of a dual cubic tetrahedron face cell as a vector a, this vector can be rotated into the observer's view coordinate system and expressed as an angular responsivity vector c (expressed in Cartesian coordinates) as:

    (52) [ c x c y c z ] = [ u x u y u z v x v y v z n x n y n z ] [ a x a y a z ] ( 9 )

    (53) The melanopic irradiance E.sub.mel that is detected by a virtual observer then becomes:
    E.sub.mel=K.sub.melΣ.sub.i=1.sup.mS.sub.iM.sub.eΔA.sub.i/π(u.sub.i.sup.2+v.sub.i.sup.2+1).sup.3/2  (10)
    where S.sub.i is the angular responsivity to melanopic irradiance in the direction of face cell i relative to the gaze direction, and K.sub.mel is a constant that is dependent on the SPD of the incident light. The summation is over all six faces of the dual cubic tetrahedron, in contrast to the summation over a single face in Equation (4).
    Spectral Responsivity

    (54) The angular responsivity function or tabulation may also incorporate the age-related melanopic action spectrum due to corneal yellowing (e.g., Turner, P. L., and M. A. Mainster. 2008. “Circadian Photoreception: Ageing and the Eye's Important Role in Systemic Health,” British Journal of Ophthalmology 92:1439-1444), as shown in FIG. 2.

    (55) Multiple Light Sources

    (56) As noted above, a virtual environment may include light sources with different SPDs with both direct and indirect light incident upon the observer. (Display screens in a virtual environment may be modeled as emissive surfaces with their own SPDs.) In addition, some light sources may be switched on and off or dimmed during the day.

    (57) For the purposes of calculating melanopic irradiance at the observer position, light sources may need to be grouped by SPD and/or lighting control channel, and the virtual environment surface patch exitances calculated separately for each group.

    (58) Daylight

    (59) The variability of daylight ranges in timescale from minutes to months. As disclosed in U.S. Pat. No. 9,955,552, historical weather records providing hourly direct normal and diffuse horizontal irradiances for a Typical Meteorological Year can be employed to calculate the ingress of direct solar and diffuse daylight into an interior environment, and from this the melanopic irradiance due to natural lighting. Sky models such as, for example, Bruneton, E. 2016. “A Qualitative and Quantitative Evaluation of 8 Clear Sky Models,” IEEE Transactions on Visualization and Computer Graphics 23(12):2641-2655, can further be employed to calculate the hourly daylight SPD for both direct sunlight and diffuse daylight, including near sunrise and sunset when the sky color changes rapidly.

    (60) Modeling dynamic daylight requires representing 4,380 radiant exitances values of each surface patch in order to represent hourly weather records in a typical meteorological year. As previously discussed, the spectral power distribution of daylight is remarkably constant, and be represented by that of CIE D65 daylight (Colorimetry, Fourth Edition, CIE 015:2018).

    (61) Colored Surfaces

    (62) The SPD of light indirectly reflected or transmitted by colored surfaces will differ from that of the light incident on the surfaces. As disclosed in U.S. Pat. No. 10,785,849 and US Patent Application 2021/0003452, it is sufficient to model the SPDs of all light sources, including daylight, as having the same SPD, such as for example that of CIE Standard Illuminant D65 (6500K) daylight. A radiosity solution is calculated for the virtual environment, which will generate a CIE XYZ tristimulus value M.sub.s for each dual cubic tetrahedron face cell. Summing these values in accordance with:
    E.sub.XYZ=Σ.sub.i=1.sup.mS.sub.iM.sub.sΔA.sub.i/π(u.sub.i.sup.2+v.sub.i.sup.2+1).sup.3/2  (11)
    will give the observer irradiance as a CIE XYZ tristimulus value, which can be transformed into a canonical spectral irradiance distribution. This SPD may then be multiplied on a per-wavelength basis by the light source group SPD and the melanopic action spectrum to calculate the melanopic irradiance E.sub.mel for each light source group.

    (63) FIG. 12 shows a flowchart presenting the calculation of melanopic irradiance from spherical irradiance for a specified observer gaze direction. In Step 1200, and in reference to US Patent Application 2021/0003452, the canonical spectral irradiance is initialized to zero.

    (64) In Step 1205, an observer gaze direction for a precalculated observer position is specified.

    (65) In Step 1210, one of the face cells from the six cubic tetrahedron faces is chosen.

    (66) In Step 1215, the selected face cell polar angles are rotated in accordance with Equation 9.

    (67) In Step 1220, the face cell angular responsivity is determined analytically or by table lookup.

    (68) In Step 1225, the face cell spherical irradiance (which may include direct spherical irradiance due to a light source) is multiplied by the face cell angular responsivity.

    (69) In Step 1230, the face cell RGB values are used with the face cell spherical irradiance value to calculate the face cell CIE XYZ tristimulus value M.sub.s.

    (70) In Step 1235, the face cell canonical spectral irradiance is calculated in accordance with US Patent Application 2021/0003452.

    (71) In Step 1240, the face cell canonical spectral irradiance is added to the canonical spectral irradiance.

    (72) In Step 1245, a decision is made as to whether there are more face cells to choose. If true, control passes to Step 1210 to choose the next face cell. If false, control passes to Step 1250.

    (73) In Step 1250, the light source SPD is specified.

    (74) In Step 1255, the canonical spectral irradiance is multiplied on a per-wavelength basis by the light source SPD.

    (75) In Step 1260, the resultant spectral irradiance is multiplied on a per-wavelength basis by the melanopic responsivity function (FIG. 1) to obtain the standard (25 year-old) melanopic irradiance.

    (76) In Step 1265, a decision is made whether the observer age is to be considered. If true, control proceed to Step 1270; otherwise the algorithm finishes.

    (77) In Step 1270, the observer age and associated pre-receptoral filtering function (FIG. 2) is optionally specified.

    (78) In Step 1275, the melanopic irradiance is multiplied by the pre-receptoral filtering function on a per-wavelength basis, following which the algorithm finishes.

    (79) While the above provides the ability to calculate melanopic irradiance for any gaze direction from a specified position, it is reasonable to assume that the gaze direction for an observer while standing will be approximately horizontal. Given this, a virtual environment can include a horizontal or approximately horizontal plane wherein the observer is likely to be standing or walking, in accordance with stochastic movement models (e.g., Hong et al. 2018). Melanopic irradiance meters can be positioned at intervals within this plane, with a small number of gaze directions (for example, at 30-degree intervals about the vertical axis) to represent precomputed melanopic irradiance values. Calculating melanopic dose would then require multiplying the irradiance values by the observer dwell time at each occupied position and gaze direction, or interpolating values between adjacent positions and gaze orientations.

    (80) Finally, circadian gating may be implemented with a time-dependent function that selectively weights the calculated melanopic dose at predetermined times of the day.

    (81) FIG. 14 shows an exemplary system for predicting and/or calculating the melanopic dose or dose rate. A processor 1424, which may include multiple constituent processors, is connected to one or more interfaces 1420, 1422 and also to one or more computer-readable memories 1426 storing a program 1430 in the form of computer-readable instructions, which, when executed by the processor, cause the system to follow the processes described in relation to FIGS. 11 and 12. The memory 1426 also stores computer-readable data 1432, which is used by the processor 1424 to run the program 1430. The data 1432 is created by the program 1430, by an external program, or both. Data 1432 includes a model of a virtual environment 1434, which corresponds to the environment for which the melanopic dose and dose rate are being calculated. User interface 1422, which may include a display, allows a user to input or create a model of an environment for which the melanopic dose and dose rate are to be calculated, and to receive an output from the system that indicates the melanopic dose rate or dose. Interface 1420 may be connected to a controller 1440 for one or more luminaires and/or fenestration devices 1444 in a building, which are controlled by the system to provide a desired level of melanopic dose rate at a given time, or a desired melanopic dose over a period of time.

    (82) The melanopic irradiance for a given observer position can also be measured in a physical environment. Referring to FIG. 13, a calibrated digital camera 1005 with high dynamic range and panorama generation capabilities is rotated about axis 1010 to generate a panoramic image 1015, wherein the height of the image corresponds to an angular range of approximately 120 degrees (50 degrees above and 70 degrees below the horizontal meridian) indoors and 90 degrees (20 degrees above and 70 degrees below) outdoors in direct sunlight.

    (83) Given a calibrated panorama image wherein the red-green-blue pixel luminances correspond to CIE XYZ tristimulus values, and if the light source SPD is known or can be measured with a spectroradiometer, the melanopic irradiance for a given view direction, taking into account the angular responsivity of the observer to melanopic irradiance and the observer age, can be calculated. This calculation can be done by, for example, by following steps similar to those described in PCT publication WO2018/227305, paragraphs [0067]-[0134].

    (84) If multiple light sources with different SPDs are present (including visual display screens and daylight), additional panoramic images can be captured and analyzed for each light source SPD. These images may take into account changes in daylight due to changing sky conditions and time of day, as well as the effects of blinds and shades. The overall melanopic irradiance is then the sum of the melanopic irradiance calculated for each light source.

    (85) As an example, FIG. 14 shows three panoramic images, wherein the first image shows a physical environment illuminated by electric lighting, the second image shows the same environment illuminated by diffuse daylight, and the third image shows the same environment illuminated by both electric lighting and diffuse daylight.

    (86) The embodiments of the invention may be varied in many ways. Such variations are not to be regarded as a departure from the scope of the invention, and all such modifications as would be obvious to one skilled in the art are intended to be included within the scope of the claims.