IDENTIFYING, REDUCING HEALTH RISKS, AND TRACKING OCCUPANCY IN A FACILITY
20230152652 · 2023-05-18
Inventors
- Nitesh Trikha (Pleasanton, CA, US)
- Rao P. Mulpuri (Saratoga, CA, US)
- Anurag GUPTA (San Jose, CA, US)
- Tanya Makker (Milpitas, CA, US)
- Emily Puth (Pleasanton, CA, US)
- Keivan Ebrahimi (Fremont, CA, US)
- Aditya Dayal (Sunnyvale, CA, US)
- Jack Kendrick Rasmus-Vorrath (Mountain House, CA, US)
- Robert Michael Martinson (Palo Alto, CA)
- Ajay MALIK (Milpitas, CA, US)
- Aaron Michael Smith (Madison, CT, US)
- Piers Iain Ivo Octavian MacNaughton (San Jose, CA, US)
Cpc classification
G09G3/3433
PHYSICS
H04K3/45
ELECTRICITY
H02J50/80
ELECTRICITY
G09G2300/026
PHYSICS
E06B9/24
FIXED CONSTRUCTIONS
G02F1/163
PHYSICS
E06B2009/2417
FIXED CONSTRUCTIONS
G09G2370/06
PHYSICS
G05B2219/25011
PHYSICS
E06B2009/2464
FIXED CONSTRUCTIONS
G05B2219/2642
PHYSICS
International classification
G02F1/163
PHYSICS
E06B9/24
FIXED CONSTRUCTIONS
G06F3/041
PHYSICS
G06F3/14
PHYSICS
G09F19/22
PHYSICS
H02J50/80
ELECTRICITY
Abstract
Disclosed herein as methods, apparatuses, non-transitory computer readable media, and systems relating to reduction and/or identification of one or more health risks in a facility. For example, by sensing a bodily characteristic of an individual in a facility, e.g., by sensing at least one environmental characteristic. For example, by sensing surface cleanliness. For example, by tracking personnel in the facility. For example, by suggesting routes in an enclosure based at least in part in personnel concentration in the facility. Disclosed herein as methods, apparatuses, non-transitory computer readable media, and systems relating to monitoring occupancy of a facility.
Claims
1. A method for tracking a plurality of individuals in a facility, comprising: (a) using a sensor system to sense a first identity having a first position at a first time and a second identity having a second position at a second time, wherein the sensor system is operatively coupled to a local network disposed in the facility, which sensor system comprises a plurality of sensors configured to sense and/or identify the first identity, the first position, the first time, the second identity, the second position, and the second time, which local network is configured to facilitate control of at least one other device of the facility; (b) tracking movement of the first identity over a time period to generate a first tracking information, and tracking movement of the second identity over the time period to generate a second tracking information; and (c) evaluating a distance from the first tracking information to the second tracking information relative to a distance threshold.
2. The method of claim 1, wherein the plurality of sensors includes a plurality of geolocation sensors that are time synchronized.
3. The method of claim 1, wherein the sensor system is comprised of a device ensemble mounted at a fixed location in the facility.
4. The method of claim 1, further comprising: (d) associating the first position and the first time with the first identity to generate a first association, and associating the second position and the second time with the second identity to generate a second association; and (e) comparing the first association with the second association to evaluate a distance from the first identity to the second identity relative to the distance threshold.
5. The method of claim 1, further comprising evaluating whether the first tracking information and the second tracking information were at a distance below the distance threshold for a cumulative time relative to a time threshold.
6-7. (canceled)
8. The method of claim 1, further comprising using the local network to transmit at least a fourth generation, or a fifth generation cellular communication and data comprising media.
9-10. (canceled)
11. The method of claim 1, further comprising using the network to control a tintable window disposed in the facility.
12-27. (canceled)
28. A method for monitoring disinfection of surfaces of a facility, comprising: (A) using a sensor system to sense a plurality of temperature samples of an object surface at a plurality of sample times, which sensor system is disposed in the facility and is operatively coupled to a local network of the facility, which local network is configured to control at least one other device of the facility that is operatively coupled to the local network; (B) comparing consecutive temperature samples of the plurality of temperature samples to generate a comparison; (C) detecting a cleaning event when the comparison indicates a temperature drop below a temperature threshold; (E) monitoring an elapsed time since a last cleaning event; and (F) generating a notification when the elapsed time exceeds a time threshold.
29. The method of claim 28, wherein the sensor system is coupled to a local network disposed in the facility in which the object surface is disposed.
30. The method of claim 29, further comprising using the local network to transmit data comprising media.
31. The method of claim 28, wherein the sensor system remotely senses the temperature samples.
32. The method of claim 28, wherein the sensor system is operatively coupled to a hierarchical control system comprising a plurality of controllers.
33-39. (canceled)
40. A method of detecting a bodily characteristic of an individual in a facility, comprising: (a) using a sensor system to sense an environmental characteristic in presence of the individual on a plurality of occasions, which sensor system is disposed in the facility and is operatively coupled to a local network that is configured to facilitate control of at least one other device of the facility; (b) analyzing (i) the plurality of environmental characteristic data samples, and (ii) a threshold indicative of abnormal bodily characteristic, to generate an analysis; and (c) using the analysis to generate a report of presence and/or absence of the indication of abnormal bodily characteristic of the individual.
41. The method of claim 40, wherein the environmental characteristic is detectibly perturbed by the presence of the individual, as compared to absence of the individual from the environment.
42. The method of claim 40, wherein collecting a plurality of environmental characteristic data samples of the individual for the plurality of occasions is to quantify a normal bodily characteristic of the individual, wherein the analysis further comprises analyzing a relative difference between a recent one of the data samples and the normal qualified, and wherein the threshold is a difference threshold.
43. The method of claim 40, wherein the sensor system comprises an electromagnetic sensor.
44. The method of claim 40, wherein the sensor system comprises a first electromagnetic sensor configured to detect a first radiation range, and a second electromagnetic sensor configured to detect a second radiation range having at least one portion that does not overlap the first radiation range.
45. The method of claim 40, further comprising focusing at least one sensor of the sensor system on one or more facial landmark features of the individual to measure the environmental characteristics.
46. The method of claim 40, further comprising focusing at least one sensor of the sensor system on depth placement of the individual to measure the environmental characteristics.
47. The method of claim 40, wherein evaluating the characteristic comprises filtering environmental characteristics attributed to the background.
48-60. (canceled)
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0072] The novel features of the invention are set forth with particularity in the appended claims. A better understanding of the features and advantages of the present invention will be obtained by reference to the following detailed description that sets forth illustrative embodiments, in which the principles of the invention are utilized, and the accompanying drawings or figures (also “Fig.” and “Figs.” herein), of which:
[0073]
[0074]
[0075]
[0076]
[0077]
[0078]
[0079]
[0080]
[0081]
[0082]
[0083]
[0084]
[0085]
[0086]
[0087]
[0088]
[0089]
[0090]
[0091]
[0092]
[0093]
[0094]
[0095]
[0096]
[0097]
[0098]
[0099]
[0100]
[0101]
[0102]
[0103] The figures and components therein may not be drawn to scale. Various components of the figures described herein may not be drawn to scale.
DETAILED DESCRIPTION
[0104] While various embodiments of the invention have been shown, and described herein, it will be obvious to those skilled in the art that such embodiments are provided by way of example only. Numerous variations, changes, and substitutions may occur to those skilled in the art without departing from the invention. It should be understood that various alternatives to the embodiments of the invention described herein might be employed.
[0105] Terms such as “a,” “an,” and “the” are not intended to refer to only a singular entity but include the general class of which a specific example may be used for illustration. The terminology herein is used to describe specific embodiments of the invention(s), but their usage does not delimit the invention(s).
[0106] When ranges are mentioned, the ranges are meant to be inclusive, unless otherwise specified. For example, a range between value 1 and value 2 is meant to be inclusive and include value 1 and value 2. The inclusive range will span any value from about value 1 to about value 2. The term “adjacent” or “adjacent to,” as used herein, includes “next to,” “adjoining,” “in contact with,” and “in proximity to.”
[0107] As used herein, including in the claims, the conjunction “and/or” in a phrase such as “including X, Y, and/or Z”, refers to in inclusion of any combination or plurality of X, Y, and Z. For example, such phrase is meant to include X. For example, such phrase is meant to include Y. For example, such phrase is meant to include Z. For example, such phrase is meant to include X and Y. For example, such phrase is meant to include X and Z. For example, such phrase is meant to include Y and Z. For example, such phrase is meant to include a plurality of Xs. For example, such phrase is meant to include a plurality of Ys. For example, such phrase is meant to include a plurality of Zs. For example, such phrase is meant to include a plurality of Xs and a plurality of Ys. For example, such phrase is meant to include a plurality of Xs and a plurality of Zs. For example, such phrase is meant to include a plurality of Ys and a plurality of Zs. For example, such phrase is meant to include a plurality of Xs and Y. For example, such phrase is meant to include a plurality of Xs and Z. For example, such phrase is meant to include a plurality of Ys and Z. For example, such phrase is meant to include X and a plurality of Ys. For example, such phrase is meant to include X and a plurality of Zs. For example, such phrase is meant to include Y and a plurality of Zs.
[0108] The term “operatively coupled” or “operatively connected” refers to a first element (e.g., mechanism) that is coupled (e.g., connected) to a second element, to allow the intended operation of the second and/or first element. The coupling may comprise physical or non-physical coupling. The non-physical coupling may comprise signal-induced coupling (e.g., wireless coupling). Coupled can include physical coupling (e.g., physically connected), or non-physical coupling (e.g., via wireless communication).
[0109] An element (e.g., mechanism) that is “configured to” perform a function includes a structural feature that causes the element to perform this function. A structural feature may include an electrical feature, such as a circuitry or a circuit element. A structural feature may include a circuitry (e.g., comprising electrical or optical circuitry). Electrical circuitry may comprise one or more wires. Optical circuitry may comprise at least one optical element (e.g., beam splitter, mirror, lens and/or optical fiber). A structural feature may include a mechanical feature. A mechanical feature may comprise a latch, a spring, a closure, a hinge, a chassis, a support, a fastener, or a cantilever, and so forth. Performing the function may comprise utilizing a logical feature. A logical feature may include programming instructions. Programming instructions may be executable by at least one processor. Programming instructions may be stored or encoded on a medium accessible by one or more processors.
[0110] In some embodiments, an enclosure comprises an area defined by at least one structure. The at least one structure may comprise at least one wall. An enclosure may comprise and/or enclose one or more sub-enclosure. The at least one wall may comprise metal (e.g., steel), clay, stone, plastic, glass, plaster (e.g., gypsum), polymer (e.g., polyurethane, styrene, or vinyl), asbestos, fiber-glass, concrete (e.g., reinforced concrete), wood, paper, or a ceramic. The at least one wall may comprise wire, bricks, blocks (e.g., cinder blocks), tile, drywall, or frame (e.g., steel frame).
[0111] In some embodiments, a plurality of devices are operatively (e.g., communicatively) coupled to a control system. The plurality of devices may be disposed in a facility (e.g., including a building and/or room). The control system may comprise the hierarchy of controllers. The devices may comprise an emitter, a sensor, or a (e.g., tintable) window (e.g., IGU). The device may be any device as disclosed herein. At least two of the plurality of devices may be of the same type. For example, two or more IGUs may be coupled to the control system. At least two of the plurality of devices may be of different types. For example, a sensor and an emitter may be coupled to the control system. At times the plurality of devices may comprise at least about 20, 50, 100, 500, 1000, 2500, 5000, 7500, 10000, 50000, 100000, or 500000 devices. The plurality of devices may be of any number between the aforementioned numbers (e.g., from about 20 devices to about 500000 devices, from about 20 devices to about 50 devices, from about 50 devices to about 500 devices, from about 500 devices to about 2500 devices, from about 1000 devices to about 5000 devices, from about 5000 devices to about 10000 devices, from about 10000 devices to about 100000 devices, or from about 100000 devices to about 500000 devices). For example, the number of windows in a floor may be at least about 5, 10, 15, 20, 25, 30, 40, or 50. The number of windows in a floor can be any number between the aforementioned numbers (e.g., from 5 to 50, from 5 to 25, or from 25 to 50). At times the devices may be in a multi-story building. At least a portion of the floors of the multi-story building may have devices controlled by the control system (e.g., at least a portion of the floors of the multi-story building may be controlled by the control system). For example, the multi-story building may have at least about 2, 8, 10, 25, 50, 80, 100, 120, 140, or 160 floors that are controlled by the control system. The number of floors (e.g., devices therein) controlled by the control system may be any number between the aforementioned numbers (e.g., from about 2 to about 50, from about 25 to about 100, or from about 80 to about 160). The floor may be of an area of at least about 150 m.sup.2, 250 m.sup.2, 500 m.sup.2, 1000 m.sup.2, 1500 m.sup.2, or 2000 square meters (m.sup.2). The floor may have an area between any of the aforementioned floor area values (e.g., from about 150 m.sup.2 to about 2000 m.sup.2, from about 150 m.sup.2 to about 500 m.sup.2, from about 250 m.sup.2 to about 1000 m.sup.2, or from about 1000 m.sup.2 to about 2000 m.sup.2).
[0112] In some embodiments, the enclosure comprises one or more openings. The one or more openings may be reversibly closable. The one or more openings may be permanently open. A fundamental length scale of the one or more openings may be smaller relative to the fundamental length scale of the wall(s) that define the enclosure. A fundamental length scale may comprise a diameter of a bounding circle, a length, a width, or a height. The fundamental length scale may be abbreviated herein as “FLS.” A surface of the one or more openings may be smaller relative to the surface the wall(s) that define the enclosure. The opening surface may be a percentage of the total surface of the wall(s). For example, the opening surface can measure about 30%, 20%, 10%, 5%, or 1% of the walls(s). The wall(s) may comprise a floor, a ceiling or a side wall. The closable opening may be closed by at least one window or door. The enclosure may be at least a portion of a facility. The enclosure may comprise at least a portion of a building. The building may be a private building and/or a commercial building. The building may comprise one or more floors. The building (e.g., floor thereof) may include at least one of: a room, hall, foyer, attic, basement, balcony (e.g., inner or outer balcony), stairwell, corridor, elevator shaft, façade, mezzanine, penthouse, garage, porch (e.g., enclosed porch), terrace (e.g., enclosed terrace), cafeteria, and/or Duct. In some embodiments, an enclosure may be stationary and/or movable (e.g., a train, a plane, a ship, a vehicle, or a rocket).
[0113] In some embodiments, the enclosure encloses an atmosphere. The atmosphere may comprise one or more gases. The gases may include inert gases (e.g., argon or nitrogen) and/or non-inert gases (e.g., oxygen or carbon dioxide). The enclosure atmosphere may resemble an atmosphere external to the enclosure (e.g., ambient atmosphere) in at least one external atmosphere characteristic that includes: temperature, relative gas content, gas type (e.g., humidity, and/or oxygen level), debris (e.g., dust and/or pollen), and/or gas velocity. The enclosure atmosphere may be different from the atmosphere external to the enclosure in at least one external atmosphere characteristic that includes: temperature, relative gas content, gas type (e.g., humidity, and/or oxygen level), debris (e.g., dust and/or pollen), and/or gas velocity. For example, the enclosure atmosphere may be less humid (e.g., drier) than the external (e.g., ambient) atmosphere. For example, the enclosure atmosphere may contain the same (e.g., or a substantially similar) oxygen-to-nitrogen ratio as the atmosphere external to the enclosure. The velocity of the gas in the enclosure may be (e.g., substantially) similar throughout the enclosure. The velocity of the gas in the enclosure may be different in different portions of the enclosure (e.g., by flowing gas through to a vent that is coupled with the enclosure).
[0114] Certain disclosed embodiments provide a network infrastructure in the enclosure (e.g., a facility such as a building). The network may be installed when an envelope of an enclosure (e.g., a building) is constructed, and/or before an interior of the enclosure is constructed. The network may comprise wiring (e.g., cables) residing in the envelope of the enclosure (e.g., external walls of the enclosure such as a building). The network infrastructure is available for various purposes such as for providing communication and/or power services. The communication services may comprise high bandwidth (e.g., wireless and/or wired) communications services. The communication services can be to occupants of a facility and/or users outside the facility (e.g., building). The network infrastructure may work in concert with, or as a partial replacement of, the infrastructure of one or more cellular carriers. The network infrastructure can be provided in a facility that includes electrically switchable windows. Examples of components of the network infrastructure include a high speed backhaul. The network infrastructure may include at least one cable, switch, physical antenna, transceivers, sensor, transmitter, receiver, radio, processor and/or controller (that may comprise a processor). The network infrastructure may be operatively coupled to, and/or include, a wireless network. The network infrastructure may comprise wiring. One or more sensors can be deployed (e.g., installed) in an environment as part of installing the network and/or after installing the network. The network may be an encrypted network. The network may comprise one or more levels of encryption.
[0115] In some embodiments, the network infrastructure is operatively coupled to one or more sensors disposed in the enclosure. The sensor(s) may be configured to detect an environmental characteristic influenced by an individual present in the enclosure, and/or any gradient of that characteristic in the environment. The sensor may be able to sense and/or identify an abnormal influence on that environmental characteristic by the individual (e.g., excessive heat, perspiration, VOC, and/or carbon dioxide emission). In some embodiment, the sensor is configured to sense and/or identify one or more bodily characteristics emitted and/or expelled (e.g., directly) from the individual (e.g., heat, perspiration (e.g., humidity), VOCs (e.g., odor), and/or carbon dioxide). The bodily characteristic may comprise sebum, or urine. The VOCs can be reaction product(s) of compounds excreted by the individual. For example, formaldehyde, or (e.g., cyclic) methylsiloxane. In some embodiment, the sensor is configured to sense and/or identify one or more bodily characteristics of the individual (e.g., directly sense the individual). The sensor may be operatively coupled to a network (e.g., the network infrastructure disclosed herein). The sensed data may be analyzed and/or recorded. The analyzed data may be utilized to provide a report and/or alert. For example, the analyzed data may send an alert if the sensed data confines to a rule. The rule may comprise if the data is outside a threshold window, inside a threshold window, above a threshold, or below a threshold. The threshold may be a value or a function. The threshold may evolve in time, e.g., as more data emerges. The rule may be predetermined, e.g., by a user, by a third party, by a global and/or jurisdictional standard, or by any combination thereof.
[0116] In some embodiments, the abnormal bodily characteristic is related to an illness. The illness may relate to an inflammation. The illness may comprise pathogen (e.g., bacterial or viral) related infections. The pathogen may cause common cold, pneumonia or flu like symptoms. The pathogen may include streptococcus. The pathogen may comprise rhino, influenza, or corona virus. The illness may relate to an epidemic. The illness may relate to a respiratory and/or inflammatory disease. The illness may relate to an epithelial tissue and/or neuronal infectious disease. The illness may relate to a severe and/or acute disease. The illness may relate to nose, ear and/or throat infections. The illness may be contagious (e.g., by individuals). The illness may infect the bronchia and/or lungs. The illness may infect one or more internal organs and/or blood. The illness may comprise pneumonia, or flu. The illness may be caused by microorganisms, e.g., that invade at least one bodily tissue. The illness may be contagious having an Ro factor of at least about 0.5, 1, 1.5, 2, 2.5, or 3.
[0117] In some embodiments, one or more sensors may transmit the analyzed data, report and/or alert. The transmission may be to an individual and/or to a user. For example, the report may be transmitted to a party of interest. The party of interest may be an establishment head, a facility owner, a user, a tested individual. The data may be saved, analyzed and/or reported in compliance with jurisdictional rules and regulations. The report, notice, alert, and/or data may be transmitted to wired and/or wireless device (e.g., cell phone). For example, an individual tested may receive a sound, vibration, picture or written notice (e.g., alert). The transmission may comprise an encoding. The encoding may comprise one signal for an abnormal characteristic (e.g., suggesting an ill individual), and a second signal for a normal characteristic (e.g., suggesting a healthy individual). For example, a cell phone of a person my vibrate once when a temperature sensor does not sense an elevated temperature for the person and vibrate twice when the temperature sensor senses an elevated temperature for the person.
[0118] In some embodiments, the sensor(s) are disposed in an environment. A sensor may be attached to (or disposed in) at least one wall, ceiling, or floor of an enclosure. In an example, the sensor or an ensemble of sensors may dispose on the floor or in the floor. The ensemble of sensors may comprise a weight sensor, a movement sensor, and accelerometer, an audio sensor, or an optical sensor. At least one sensor in the ensemble may be able to sense and/or identify a presence of an individual, e.g., by sensing a person's weight, by sensing a person's sound, by sensing a person's obstruction of light, by sensing an approaching object associated with the person. For example, when an individual resides on a platform coupled to a weight sensor, the weight sensor may be able to sense and/or identify a weight associated with an individual. For example, when an individual obstructs a light ray (e.g., IR light ray) sensed by an optical sensor, the optical sensor may be able to sense and/or identify the obstruction of light. A requested position of the individual (e.g., testing station) may be marked on the floor (e.g., in a form of a shape such as a circle, square, or the letter X). The sensor(s) may be mounted on the door (e.g., door frame), or on the window (e.g., window frame). The enclosure may facilitate testing of a single individual or a plurality of individuals. There may be a single testing station (e.g., a kiosk) disposed in an enclosure, or a plurality of testing stations. Testing stations of the plurality of testing stations may be sufficiently distant to prevent contamination between individuals in the same room (e.g., prevents inter-contamination of individuals). The distance may be predefined and/or adjusted and/or defined per jurisdictional, global, and/or third party standards. The person may freely walk over and/or past the designated testing area or stay there for a time. The time in which the individual may be required to remain (e.g., substantially) stationary on the testing area may be at most about 1 second (sec), 3 sec, 5 sec, 10 sec, 30 sec, or 60 sec. The time in which the individual may be required to remain (e.g., substantially) stationary on the testing area may be any value between the aforementioned values (e.g., from about 1 sec to about 60 sec, from about 1 sec, to about 5 sec, from about 5 sec to about 30 sec, or from about 30 sec to about 60 sec). An individual may pass by the sensor(s) and be tested on every passing. This may allow for automatic testing of an individual multiple times in a single day, depending on how many times the individual passes by the testing station (e.g., by the sensor(s)). A plurality of test results may increase confidence in the results, reduce error values, and/or enable identification of relative changes between test results obtained at different times. If any of the sensors have a high fidelity, the sensor(s) may be able to provide information of trends of the characteristic in the individual in absolute and/or relative terms, e.g., over periods of time such as at least an hour, a day, a week, a month, or a year. If a sensor is uncalibrated or uncalibratable, and the sensor results are consistent over periods of usage, then the sensor may be able to provide information of trends of the characteristic in the individual in relative terms, e.g., over periods of time such as at least an hour, a day, a week, a month, or a year.
[0119]
[0120] The testing station (e.g., kiosk) may contain one or more sensors, e.g., that perform the environmental position and/or detect the presence of the individual. The testing station may be devoid of at least a sensor that perform the environmental position. The testing station may be devoid of at least a sensor that detect the presence of the individual. For example, a weight sensor may detect the presence of the individual and be disposed in testing platform. The temperature sensor that may detect the temperature of the environment in the enclosure may be disposed in the testing platform, or elsewhere in the enclosure. For example, an optical sensor (e.g., sensor sensitive to electromagnetic radiation such as in the infrared and/or visible range) may detect the presence of the individual and be disposed in the enclosure outside of the platform (e.g., not in the testing platform). The platform may be devoid of the one or more sensors. The testing platform may direct the individual where to position in order for the testing to be (e.g., optimally) performed. The ingress and/or ingress to the enclosure (e.g., doors) may be closed during testing. The sensor(s) may be disposed at a fixture of a facility (e.g., testing station, entrance hall, or conference room). The fixture may comprise a framing, a wall, a ceiling, or a floor. The framing may comprise a door framing or a window framing.
[0121] In some embodiments, an occupancy sensor is operatively coupled to the network. The occupancy sensor may be configured to sense and/or identify motion of, electromagnetic radiation associated with, and/or identification tag of, an occupant in the facility (or entering the facility). For example, the sensor may comprise a visible or an IR sensor. At times, the occupancy sensor may be unable to identify an identity of the occupant. The IR sensor may detect heat signature characteristic of individuals. The IR sensor may be part of a sensor system (e.g., a device ensemble). Data of the IR sensor may be integrated with visible data and/or motion data. The sensor system may be operatively coupled to the network and/or to the control system of the facility. The occupancy sensor (e.g., as part of the sensing system) may be utilized to determine density of individuals in an enclosure of the facility in which the occupancy sensor(s) is disposed and/or user of the enclosure by occupants (e.g., over time). The occupancy sensor may be operatively coupled to at least one processor. The occupancy sensor may be configured to detect movement of occupants. For example, the occupancy sensor may detect electromagnetic radiation emitted by and/or reflected from an occupant (e.g., IR radiation such as forming a heat signature of the individual), which detention may occur overtime. In some embodiments, using the occupancy sensor (e.g., as part of a sensor system such as a device ensemble) comprises measuring at a rate of at least about every 2 seconds (sec), 1 sec, 0.5 sec, 0.25 sec, 0.2 sec, 0.15 sec, 0.1 sec, 0.05 sec, or 0.025 sec. The occupancy sensor may comprise a camera. The camera may be configured to take at least about 30 frames per second (frm/sec), 20 frm/sec, 10 frm/sec, 8 frm/sec, 6 frm/sec, 4 frm/sec, or 2 frm/sec. The frequency of sensing (e.g., the number of measurements per second taken, such as the number of frames per seconds taken) may be adjusted (e.g., manually and/or automatically using at least one controller, e.g., as part of the control system). The camera may comprise an IR and/or a visible camera. The sensor system may be configured to detect a location of the occupant, e.g., relative to the facility, relative to the enclosure in which it is disposed, and/or in absolute coordinates (e.g., using a GPS or any other anchored geo-location technology). Detection of the occupant overtime and space may be utilized to determine a movement (e.g., including a movement direction) of the occupant. The movement direction of the occupant may comprise utilizing a tracking algorithm. Determination of occupancy and/or movement of the occupant may be determined to an accuracy of at least about 80%, 85%, 90%, 95%, or 99%. The tracking algorithm may be a tracking algorithms such as is disclosed in “Simple object tracking with OpenCV” by Adrian Rosebrock published on Jul. 23, 2018 and available at https://www.pyimagesearch.com/2018/07/23/simple-object-tracking-with-opencv, which is incorporated herein by reference in its entirety.
[0122]
[0123]
[0124] In some embodiments, the control system is operatively (e.g., communicatively) coupled to an ensemble of devices (e.g., sensors and/or emitters). The ensemble facilitates the control of the environment and/or the alert. The control may utilize a control scheme such as feedback control, or any other control scheme delineated herein. The ensemble may comprise at least one sensor configured to sense and/or identify electromagnetic radiation. The electromagnetic radiation may be (humanly) visible, infrared (IR), or ultraviolet (UV) radiation. The at least one sensor may comprise an array of sensors. For example, the ensemble may comprise an IR sensor array (e.g., a far infrared thermal array such as the one by Melexis). The IR sensor array may have a resolution of at least 32×24 pixels, or 640×480 pixels. The IR sensor may be coupled to a digital interface. The ensemble may comprise an IR camera. The ensemble may comprise a sound detector. The ensemble may comprise a microphone. The ensemble may comprise any sensor and/or emitter disclosed herein. The ensemble may include CO.sub.2, VOC, temperature, humidity, electromagnetic light, pressure, and/or noise sensors. The sensor may comprise a gesture sensor (e.g., RGB gesture sensor), an acetometer, or a sound sensor. The sounds sensor may comprise an audio decibel level detector. The sensor may comprise a meter driver. The ensemble may include a microphone and/or a processor. The ensemble may comprise a camera (e.g., a 4K pixel camera), a UWB sensor and/or emitter, a UHF sensor and/or emitter, a Bluetooth (abbreviated herein as “BLE”) sensor and/or emitter, a processor. The camera may have any camera resolution disclosed herein. One or more of the devices (e.g., sensors) can be integrated on a chip. The sensor ensemble may be utilized to determine presence of occupants in an enclosure, their number and/or identity (e.g., using the camera). The sensor ensemble may be utilized to control (e.g., monitor and/or adjust) one or more environmental characteristics in the enclosure environment.
[0125] The sensor(s) attached to the opening (e.g., door and/or window) by which the individual passes, may contain one or more sensors. The one or more sensors may perform the environmental position and/or detect the presence of the individual. The sensor(s) attached to the opening may be devoid of at least a sensor that perform the environmental position. The sensor(s) attached to the opening may be devoid of at least a sensor that detect the presence of the individual. For example, an environmental characteristic influenced by the individual (e.g., temperature) may be disposed in the sensor(s) attached to the opening or disposed elsewhere in the enclosure. For example, a sensor that senses the presence of the individual (e.g., optical sensor) may be attached to the opening, or disposed elsewhere in the enclosure. The system detecting an environmental characteristic(s) influenced by an individual may be devoid of any sensor that does not detect environmental characteristic(s). A sensor that does not detect environmental characteristic(s) may detect a presence of an individual (e.g., a weight sensor, or a camera). In some embodiments, a sensor that detects an environmental characteristic also detects the presence of an individual (e.g., infrared sensor, humidity sensor, or carbon dioxide sensor). In some embodiments, one or more openings in the enclosure remain open at least in part during the testing. For example, a door may be kept open while the sensors sense the temperature of an environment of the opening (e.g., and thus any perturbation thereof). The perturbation may be caused by an individual. An individual may have a typical manner of perturbing an environmental characteristic (e.g., within normal and/or abnormal range). In some embodiments, sensor(s) detecting whether a person influences an environmental change indicative of normal and/or abnormal personal characteristics is sufficient, e.g., and there is no additional requirement for a sensor dedicated to indicating presence or absence of the individual. For example, high fidelity (e.g., high resolution) sensors measuring a personal characteristic may be sufficient. For example, a sensor that (e.g., markedly) perturbs an ambient environmental characteristic may be sufficient. A high fidelity sensor may have more than one single pixel detector. A high fidelity sensor may comprise a plurality (e.g., an array) of single pixel detectors. A high fidelity sensor may be configured to detect a span of values of the detectable properties. For example, an amplitude span (e.g., range). For example, a wavelength span (e.g., including a plurality of wavelengths). For example, a thermal detector may be an optical detector (e.g., an infrared (IR) detector), and may be able to detect a plurality of IR wavelengths (e.g., within a wavelength range). A marked perturbation may have a high signal to noise ratio that is detectibly different from a signal at ambient conditions. In some embodiments, sensor(s) detecting the environmental characteristic (e.g., temperature) form an array of sensors. The array of sensors may be included in an ensemble. The ensemble and/or sensor(s) therein may have a small FLS, small form factor (e.g., volume to height ratio), may be of low cost, may be Bluetooth enabled (e.g., comprise a Bluetooth adaptor), and/or require low power. The small FLS may be of at most about 20 millimeters (mm), 10 mm, 8 mm, 6 mm, 5 mm, 4 mm, 3 mm, 2 mm or 1 mm. The small FLS may be of any value between the aforementioned values (e.g., from about 20 mm to about 1 mm, from about 20 mm to about 8 mm, or from about 8 mm to about 1 mm). The low power required by the ensemble and/or sensor(s) therein may be of at most about 1500 microwatts (W), 1000 W, 800 W, 500 W, or 250 W. The low power required by the ensemble and/or sensor(s) therein may be between any of the abovementioned power values (e.g., from about 1500 W to about 250 W, from about 1500 W to about 500 W, or from about 500 W to about 250 W).
[0126] In some embodiments, the testing station includes a barrier. The barrier may selectively allow an individual to pass therethrough. The barrier may comprise one or more doors that (i) hinder an occupant from walking therethrough (e.g., block passage of the occupant) when the door(s) are closed or (ii) allow the occupant to walk past the door(s) when they are open. The doors may comprise a transparent material or a non-transparent material. The door may comprise glass, polymeric, or metallic material. The door may swivel about a hinge. The door may slide, e.g., on bearings. The door may be a rolling door, e.g., about an axle. The door may be operatively coupled (e.g., connected) to one or more posts. The door may be operatively coupled (e.g., connected) to a separator. The separator may facilitate separation of occupants. The separator may comprise a cavity. The cavity may comprise one or more circuitry. The door may automatically and/or reversibly open and close. Closure and opening of the door may be controller by one or more controllers (e.g., as disclosed herein). The door may be operatively coupled to an actuator configured to facilitate opening and/or closing the door, e.g., upon signal received from the controller(s). The actuator may be operatively coupled to the controller(s), e.g., via a network. The separator may or may not comprise at least one controller configured to control the opening and/or closing of the door. The separator may comprise the actuator. The barrier apparatus may include a security gate, a bar, a stile such as a turnstile (e.g., tripod turnstile), and/or a swing gate. The turnstiles may be instead or in addition to the door. The bar and/or stile may comprise any material disclosed herein (e.g., door material) The barrier may comprise an access control barrier. The door may be a fraction of an average individual height. The fraction may be at least about 0.25, 0.3, 0.4, 0.5, 0.75, 1, 1.25, or 1.5 times a height of an average individual.
[0127]
[0128] In some embodiments, the testing station includes one or more frames. The frames may hold an object. The frames and/or framed object may be modular. For example, the relative location of the frames and/or framed object may be altered. The framed object may comprise a window (e.g., tintable window), a board, and/or a display construct. The display construct may comprise a monitor (e.g., a computer or a television monitor). The display construct may comprise a light emitting diode (LED) array such as an organic LED array. The Organic LED array may be at least partially transparent (T-OLED). Examples of display construct, their operation and control can be found in U.S. Provisional Patent Application Ser. No. 62/975,706, filed on Feb. 12, 2020, titled “TANDEM VISION WINDOW AND MEDIA DISPLAY,” that is incorporated herein by reference in its entirety. The display construct may be disposed in the frame as a stand-alone framed object. The display construct may be disposed on a window (e.g., a tintable window). The framed object may comprise one or more dispensing and/or sanitation related apparatuses. The dispensing apparatuses may be configured to dispense a glove, a mask, a sanitation liquid, a badge, paper (e.g., tissue). The dispensing apparatus may be a printer (e.g., dispensing printed stickers or papers). The frame may comprise wiring. The frame may comprise one or more devices including a sensor, an emitter, an antenna, a controller, a circuitry, and/or a power source. The testing station may be operatively (e.g., communicatively) coupled to the network (e.g., and to one or more controllers). The display construct (e.g., OLED display) may display messages to an individual seeking to enter the facility. The display construct may display an image of a receptionist. The receptionist may be stationed in the facility or outside of the facility. The receptionist image may be of a live receptionist (e.g., interacting with the individual via video conferencing). The receptionist image may be of an animation. The testing station may comprise a signaling image. The signaling image may be stationary or non-stationary. The signaling image may be changing or non-changing over time. The signaling image may be permanent, or non-permanent (e.g., ephemeral). The signaling image may be transparent or non-transparent. The signaling image may be affixed (e.g., to a fixture of the enclosure such as a floor), or non-affixed. For example, the signaling image may comprise a sticker. The signaling image comprise a painting. The signaling image may comprise carving, embossing, or embroidery. The signaling image may be part of a fabric, e.g., a carpet. The signaling image may be a projection of light. The projection of flight may be changing or non-changing over time. The testing framing may comprise a projector projecting lighting on the individual and/or on the floor to form the signaling image. The projected lighting may signal an individual where to stand (e.g., a circle may be projected signaling the individual to stand in the circle). The projected lighting may signal an individual in what direction to stand (e.g., the projected lighting may comprise footprints of a stationary individual disposed in a direction, signaling the individual to stand in the direction). The projected lighting may signal an individual in what direction to walk (e.g., the projected lighting may comprise footprints of a walking individual disposed in a direction, signaling the individual to walk in the direction). The projection may be in a color. The projection may be a white projection. The color may signal if the individual is ready to proceed to the next stage in the admission process (e.g., a green color indicating that the individual is free to enter the facility). The color may signal if the individual is not ready to proceed to the next stage in the admission process (e.g., a red color indicating that the individual is unable to enter the facility). The color may signal an error or malfunction (e.g., an orange or yellow color indicating that there is an error in the process and/or in one of the components of the testing station).
[0129]
[0130]
[0131] In some embodiments, the framing apparatus (e.g., framing system) and the barrier apparatus are disposed in a reception area of the facility. The framing apparatus may be disposed in close proximity (e.g., contacting or non-contacting) the barrier apparatus. For example, the framing apparatus may be disposed in a walking distance from the barrier apparatus. The walking distance may comprise at most about 10, 8, 5, 3, or 2 average steps of an individual.
[0132] The testing station may comprise a framing apparatus comprising framing (e.g., including a mullion and a transom) that frames one or more framed objects. An external surface of the framed object may comprise a smooth or a rough portion. The rough portion may be due to embossing, sanding, or scratching of the surface. The rough portion may comprise intrusions, extrusions, or an applicant. The applicant may comprise a mesh or a cloth. The rough portion may comprise a regular pattern or an irregular pattern. The framed object may be hollow or non-hollow. The framed object may comprise an internal cavity. The internal cavity may include at least one device (e.g., sensor, emitter, controller, circuitry (e.g., processor), computer, memory, radar, or antenna). The internal cavity may comprise a transceiver. The internal cavity may comprise a circuitry. The internal cavity may comprise wiring. The internal cavity may have a Bluetooth, Global Positioning System (GPS), UHF, and/or ultrawideband (UWB) device. the internal cavity may comprise a modem. The framed object may comprise one or more slits and/or holes. The framed object may comprise on or more lightings (e.g., disposed in the compartment). The slit and/or hole may facilitate transmission of environmental characteristic from the ambient environment to the sensor(s) disposed in the separator and/or emissions (e.g., sound and/or light) to travel from the separator interior into the ambient environment.
[0133] In some embodiments, the framing apparatus or the barrier apparatus are disposed in a reception area of the facility. The entrance is to the facility may comprise one or more signaling images, e.g., signaling individuals where to stand. The signaling image may signal individuals to stand at a prescribed distance from each other. The signaling image may comprise an inner shape of allowed standing area and a rim of non-allowed standing area. The signaling image may comprise a shape (e.g., a geometric shape). The shape may comprise an ellipsoid (e.g., a circle), or a polygon. The polygon may comprise a triangle (e.g., equilateral), rectangle (e.g., square), pentagon, hexagon, heptagon, or octagon. The polygon may be a right polygon. The rim may be of the same shape as the inner shape. The rim may be of a different shape from the inner shape.
[0134] In some embodiments, the dispensing apparatus is operatively coupled to the network. The dispensing apparatus may be manually or automatically controlled. For example, the dispensing apparatus may be controlled by the controller(s). The dispensing apparatus may be operatively coupled (e.g., wired and/or wirelessly) to the network. The dispensing apparatus may be operatively (e.g., communicatively) coupled to the barrier apparatus and/or to the framing apparatus. The dispensing apparatus may be physically coupled (e.g., using an intermediate board) to the framing apparatus and/or barrier apparatus, e.g., for support and/or stability. The dispensing apparatus may be uncoupled and/or not-connected with the network.
[0135]
[0136] In some embodiments, a testing station is disposed in an admission area of a facility. The admission may comprise one or more doors. The doors may be part of the testing station or not part of the testing station. The entry way may be restricting and allow a small number of individuals (e.g., at most 5, 4, 3, 2, or 1) to pass therethrough, or may be less restricting and allow a larger number of individuals to pass therethrough. The testing station may include one or more barrier apparatuses and/or one or more framing apparatuses. At least two barrier apparatuses in the testing station may be different from each other. At least two barrier apparatuses in the testing station may be similar to each other (e.g., 425 and 426 in
[0137]
[0138]
[0139] In some embodiments, the reception area may comprise components of the testing station (e.g., a barrier apparatus and a framing apparatus). At least two of the components may be the same. At least two of the components may be different. The two components may be disposed (e.g., substantially) parallel to each other. The two components may be disposed at an angle from each other. For example, the two components may form an angle with each other, which angle may be at least about 30 degrees (°), 60°, 90°, 120°, or 180°. The angle may have any value between the aforementioned values.
[0140] In some embodiments, the testing station is disposed in an enclosure comprising a secondary (e.g., manual and/or more rigorous) testing station.
[0141] In some embodiments, a framing apparatus forms a pair with a barrier apparatus. The framing apparatus and/or barrier apparatus may be coupled to a network (e.g., via wired and/or wireless communication). The framing apparatus and barrier apparatus may or may not be directly operatively (e.g., communicatively) coupled to each other. The framing apparatus and barrier apparatus may be operatively (e.g., communicatively) coupled to each other via the network. The framing apparatus and/or barrier apparatus may comprise at least one controller (e.g., disposed therein or thereon). The framing apparatus and barrier apparatus may be operatively (e.g., communicatively) coupled to at least one controller (e.g., via the network).
[0142] In some embodiments, an individual entering the facility undergoes one or more reception procedures (e.g., operations). The reception operations may relate to entry to an admission process, equipment dispensing, bodily characteristics (e.g., sanitation), barrier, and/or to a secondary check. If the individual is a non-employee, then the operations involve a host escorting the individual into the facility and/or signature of an agreement (e.g., non-disclosure agreement). The operations may be at any order.
[0143]
[0144]
[0145] In some embodiments, the testing system is operatively coupled to one or more controllers. The controllers are configured to control one or more devices of the facility. The analysis of the testing system may be carried out at least in part by the one or more controllers that control the facility (e.g., that may comprise or may operatively couple to a building management system).
[0146] In some embodiments, one or more sensors are attached to a mobile traveler. The traveler can be animate or inanimate. For example, the senor(s) can be carried by a person, that walks about a facility. For example, the sensor(s) can be carried by a transitory robot (e.g., self-propelled vehicle). For example, an airborne vehicle (e.g., a drone) or a ground transported vehicle (e.g., a wheeled vehicle). A location of the traveler may be known. The traveler and/or tested individual may identify its location. The location of the traveler and/or tested individual may be identified at least in part via electromagnetic radiation related technology, for example, via satellite (e.g., GPS), Bluetooth, UHF and/or UWB technology. The location of the traveler may be identified at least in part via at least one sensor in the facility (e.g., a radar, a moment sensor, an antenna, and/or an optical sensor (e.g., a camera such as a video camera)). The location of the traveler and/or tested individual may be identified at least in part via the mobile device (e.g., cellular phone) that the individual and/or traveler is carrying.
[0147] In some embodiments, an environmental property excreted and/or influenced by an individual has a propagation pattern in the environment. The propagation pattern may have a distinct and/or identified pattern, e.g., that can be measured by one or more sensors. For example, a heat emitted by an individual has a diffusion pattern in an environment atmosphere. For example, a carbon dioxide emission by an individual, has an emission pattern in an environment. For example, the carbon dioxide can be more pronounced around the mouth and the nose of the individual, e.g., and the expulsion rate, direction, pressure, and/or speed can be measured. In the example shown in
[0148] The enclosure may have an ambient pressure and/or temperature. The enclosure may have a pressure different from an ambient pressure (e.g., lower and/or higher than the ambient pressure). The enclosure may have a temperature different from an ambient temperature (e.g., lower and/or higher than the ambient temperature). The enclosure be a specialized enclosure (e.g., configured to hold the pressure and/or temperature different from ambient temperature). The ambient pressure may be one atmosphere. The ambient temperature may be about 25° C. The ambient pressure and atmosphere may be standard pressure and atmosphere.
[0149] In some examples, data from a plurality of sensors (e.g., of the same type or of different types) can be utilized to assess the environment and individual(s) effect on the environment. Such assessment may shed light on the wellbeing of the individual (e.g., having normal or abnormal bodily characteristics). For example, temperature, humidity, and/or carbon dioxide patterns in an environment may be utilized to locate and/or assess the wellbeing of individual(s). data from the plurality of (e.g., different type of) sensors can be correlated (e.g., cross correlated).
[0150] In some embodiments, owners and/or users of the data may take actions if the individual exhibits one or more abnormal characteristics. For example, they may initiate medical treatment of the individual, quarantine the individual, and/or take steps to remove the individual (e.g., to reduce risk of harming others). In some embodiments, the control system of the facility may initiate an action. For example, the control system may after, or direct alteration of, the status of one or more devices operatively coupled to a building management system. For example, the control system may after, or direct alteration of, the status of an HVAC system, a lightening system, or a buzzer. For example, lighting may flash, buzzer will buzz, or ventilation will increase, in response to the result.
[0151] In some embodiments, an enclosure includes one or more sensors. The sensor may facilitate controlling the environment of the enclosure such that occupants of the enclosure may have an environment that is more comfortable, delightful, beautiful, healthy, productive (e.g., in terms of occupant performance), easer to live (e.g., work) in, or any combination thereof. The sensor(s) may be configured as low or high resolution sensors. Sensor may provide on/off indications of the occurrence and/or presence of a particular environmental event (e.g., one pixel sensors). In some embodiments, the accuracy and/or resolution of a sensor may be improved via artificial intelligence analysis of its measurements. Examples of artificial intelligence techniques that may be used include: reactive, limited memory, theory of mind, and/or self-aware techniques know to those skilled in the art). Sensors may be configured to process, measure, analyze, detect and/or react to one or more of: data, temperature, humidity, sound, force, pressure, electromagnetic waves, position, distance, movement, flow, acceleration, speed, vibration, dust, light, glare, color, gas(es), and/or other aspects (e.g., characteristics) of an environment (e.g., of an enclosure). The gases may include volatile organic compounds (VOCs). The gases may include carbon monoxide, carbon dioxide, water vapor (e.g., humidity), oxygen, radon, and/or hydrogen sulfide.
[0152] In some embodiments, processing sensor data comprises performing sensor data analysis. The sensor data analysis may comprise at least one rational decision making process, and/or learning. The sensor data analysis may be utilized to adjust and environment, e.g., by adjusting one or more components that affect the environment of the enclosure. The data analysis may be performed by a machine based system (e.g., a circuitry). The circuitry may be of a processor. The sensor data analysis may utilize artificial intelligence. The sensor data analysis may rely on one or more models (e.g., mathematical models). In some embodiments, the sensor data analysis comprises linear regression, least squares fit, Gaussian process regression, kernel regression, nonparametric multiplicative regression (NPMR), regression trees, local regression, semiparametric regression, isotonic regression, multivariate adaptive regression splines (MARS), logistic regression, robust regression, polynomial regression, stepwise regression, ridge regression, lasso regression, elasticnet regression, principal component analysis (PCA), singular value decomposition, fuzzy measure theory, Borel measure, Han measure, risk-neutral measure, Lebesgue measure, group method of data handling (GMDH), Naive Bayes classifiers, k-nearest neighbors algorithm (k-NN), support vector machines (SVMs), neural networks, support vector machines, classification and regression trees (CART), random forest, gradient boosting, or generalized linear model (GLM) technique.
[0153] Sensors of a sensor ensemble may collaborate with one another. A sensor of one type may have a correlation with at least one other type of sensor. A situation in an enclosure may affect one or more of different sensors. Sensor readings of the one or more different may be correlated and/or affected by the situation. The correlations may be predetermined. The correlations may be determined over a period of time (e.g., using a learning process). The period of time may be predetermined. The period of time may have a cutoff value. The cutoff value may consider an error threshold (e.g., percentage value) between a predictive sensor data and a measured sensor data, e.g., in similar situation(s). The time may be ongoing. The correlation may be derived from a learning set (also referred to herein as “training set”). The learning set may comprise, and/or may be derived from, real time observations in the enclosure. The observations may include data collection (e.g., from sensor(s)). The learning set may comprise sensor(s) data from a similar enclosure. The learning set may comprise third party data set (e.g., of sensor(s) data). The learning set may derive from simulation, e.g., of one or more environmental conditions affecting the enclosure. The learning set may compose detected (e.g., historic) signal data to which one or more types of noise were added. The correlation may utilize historic data, third party data, and/or real time (e.g., sensor) data. The correlation between two sensor types may be assigned a value. The value may be a relative value (e.g., strong correlation, medium correlation, or weak correlation). The learning set that is not derived from real-time measurements, may serve as a benchmark (e.g., baseline) to initiate operations of the sensors and/or various components that affect the environment (e.g., HVAC system, and/or tinting windows). Real time sensor data may supplement the learning set, e.g., on an ongoing basis or for a defined time period. The (e.g., supplemented) learning set may increase in size during deployment of the sensors in the environment. The initial learning set may increase in size, e.g., with inclusion of additional (i) real time measurements, (ii) sensor data from other (e.g., similar) enclosures, (iii) third party data, (iv) other and/or updated simulation.
[0154] In some embodiments, data from sensors may be correlated. Once a correlation between two or more sensor types is established, a deviation from the correlation (e.g., from the correlation value) may indicate an irregular situation and/or malfunction of a sensor of the correlating sensors. The malfunction may include a slippage of a calibration. The malfunction may indicate a requirement for re-calibration of the sensor. A malfunction may comprise complete failure of the sensor. In an example, a movement sensor may collaborate with a carbon dioxide sensor. In an example, responsive to a movement sensor detecting movement of one or more individuals in an enclosure, a carbon dioxide sensor may be activated to begin taking carbon dioxide measurements. An increase in movement in an enclosure, may be correlated with increased levels of carbon dioxide. In another example, a motion sensor detecting individuals in an enclosure may be correlated with an increase in noise detected by a noise sensor in the enclosure. In some embodiments, detection by a first type of sensor that is not accompanied by detection by a second type of sensor may result in a sensor posting an error message. For example, if a motion sensor detects numerous individuals in an enclosure, without an increase in carbon dioxide and/or noise, the carbon dioxide sensor and/or the noise sensor may be identified as having failed or as having an erroneous output. An error message may be posted. A first plurality of different correlating sensors in a first ensemble may include one sensor of a first type, and a second plurality of sensors of different types. If the second plurality of sensors indicate a correlation, and the one sensor indicates a reading different from the correlation, there is an increased likelihood that the one sensor malfunctions. If the first plurality of sensors in the first ensemble detect a first correlation, and a third plurality of correlating sensors in a second ensemble detect a second correlation different from the first correlation, there is an increased likelihood that the situation to which the first ensemble of sensors is exposed to is different from the situation to which the third ensemble of sensors are exposed to.
[0155] Sensors of a sensor ensemble may collaborate with one another. The collaboration may comprise considering sensor data of another sensor (e.g., of a different type) in the ensemble. The collaboration may comprise trends projected by the other sensor (e.g., type) in the ensemble. The collaboration may comprise trends projected by data relating to another sensor (e.g., type) in the ensemble. The other sensor data can be derived from the other sensor in the ensemble, from sensors of the same type in other ensembles, or from data of the type collected by the other sensor in the ensemble, which data does not derive from the other sensor. For example, a first ensemble may include a pressure sensor and a temperature sensor. The collaboration between the pressure sensor and the temperature sensor may comprise considering pressure sensor data while analyzing and/or projecting temperature data of the temperature sensor in the first ensemble. The pressure data may be (i) of a pressure sensor in the first ensemble, (ii) of pressure sensor(s) in one or more other ensembles, (iii) pressure data of other sensor(s) and/or (iv) pressure data of a third party.
[0156] In some embodiments, sensor ensembles, are distributed throughout an enclosure. Sensors of a same type may be dispersed in an enclosure, e.g., to allow measurement of environmental parameters at various locations of an enclosure. Sensors of the same type may measure a gradient along one or more dimensions of an enclosure. A gradient may include a temperature gradient, an ambient noise gradient, or any other variation (e.g., increase or decrease) in a measured parameter as a function of location from a point. A gradient may be utilized in determining that a sensor is providing erroneous measurement (e.g., the sensor has a failure).
[0157] In another example of a temperature gradient, a temperature sensor installed near a window may measure increased temperature fluctuations with respect to temperature fluctuations measured by a temperature sensor installed at a location opposite the window. A sensor installed near a midpoint between the window and the location opposite the window may measure temperature fluctuations in between those measured near a window with respect to those measured at the location opposite the window. In an example, an ambient noise sensor installed near an air conditioning (or near a heating vent) may measure greater ambient noise than an ambient noise sensor installed away from the air conditioning or heating vent.
[0158] In some embodiments, a sensor of a first type cooperates with a sensor of a second type. In an example, an infrared radiation sensor may cooperate with a temperature sensor. Cooperation among sensor types may comprise establishing a correlation (e.g., negative or positive) among readings from sensors of the same type or of differing types. For example, an infrared radiation sensor measuring an increase in infrared energy may be accompanied by (e.g., positively correlated to) an increase in measured temperature. A decrease in measured infrared radiation may be accompanied by a decrease in measured temperature. In an example, an infrared radiation sensor measuring an increase in infrared energy that is not accompanied by a measurable increase in temperature, may indicate failure or degradation in operation of a temperature sensor.
[0159] In some embodiments, one or more sensors are included in an enclosure. For example, an enclosure may include at least 1, 2, 4, 5, 8, 10, 20, 50, or 500 sensors. The enclosure may include a number of sensors in a range between any of the aforementioned values (e.g., from about 1 to about 1000, from about 1 to about 500, or from about 500 to about 1000). The sensor may be of any type. For example, the sensor may be configured (e.g., and/or designed) to measure concentration of a gas (e.g., carbon monoxide, carbon dioxide, hydrogen sulfide, volatile organic chemicals, or radon). For example, the sensor may be configured (e.g., and/or designed) to measure ambient noise. For example, the sensor may be configured (e.g., and/or designed) to measure electromagnetic radiation (e.g., RF, microwave, infrared, visible light, and/or ultraviolet radiation). For example, the sensor may be configured (e.g., and/or designed) to measure security-related parameters, such as (e.g., glass) breakage and/or unauthorized presence of personnel in a restricted area. Sensors may cooperate with one or more (e.g., active) devices, such as a radar or lidar. The devices may operate to detect physical size of an enclosure, personnel present in an enclosure, stationary objects in an enclosure and/or moving objects in an enclosure.
[0160] In some embodiments, the sensor is operatively coupled to at least one controller. The coupling may comprise a communication link. A communications link (e.g.,
[0161] In some embodiments, the enclosure is a facility (e.g., building). The enclosure may comprise a wall, a door, or a window. In some embodiments, at least two enclosures of a plurality of enclosures are disposed in the facility. In some embodiments, at least two enclosures of a plurality of enclosures are disposed different facilities. The different facilities may be a campus (e.g. and belong to the same entity). At least two of the plurality of enclosures may reside in the same floor of the facility. At least two of the plurality of enclosures may reside in different floors of the facility. Enclosures of
[0162] In an example, for gas sensors disposed in a room (e.g., in an office environment), a relevant parameter may correspond to gas (e.g., CO.sub.2) levels, where desired levels are typically in a range of about 1000 ppm or less. In an example, a CO.sub.2 sensor may determine that self-calibration should occur during a time window where CO.sub.2 levels are minimal such as when no occupants are in the vicinity of the sensor (e.g. see CO.sub.2 levels before 18000 seconds in
[0163] Positional and/or stationary characteristics (e.g., placement of walls and/or windows) of the enclosure may be utilized in measuring the characteristics of a give environment. The positional and/or stationary characteristics of the enclosure may be derived independently (e.g., from 3.sup.rd party data and/or from non-sensor data). The positional and/or stationary characteristics of the enclosure may be derived using data from the one or more sensors disposed in the environment. When the environment is minimally disturbed with respect to the measured environmental characteristic (e.g., when no one is present in an environment, and/or when the environment is quiet), some sensor data may be used to sense and/or identify position of (e.g., stationary and/or non-stationary) objects to determine the environment. Determining the position of objects comprises determining an (e.g., human) occupancy in the environment. Distance and/or location related measurements may utilize sensor(s) such as radar and/o ultrasonic sensors. Distance and location related measurements may derive from sensors that to not traditionally correlated to location and/or distance. Objects disposed in, or that are part of, an enclosure may have distinct sensor signatures. For example, location of people in the enclosure may correlate to distinct temperature, humidity and/or CO.sub.2 signatures. For example, location of a wall may correlate to an abrupt change in the distribution of temperature, humidity and/or CO.sub.2 in the enclosure. For example, location of a window or door (whether open or closed) may correlate to a change in the distribution of temperature, humidity and/or CO.sub.2 next to the window or door. The one or more sensors in the enclosure may monitor any environmental changes and/or correlates such changes to changes in subsequently monitored values. In some cases, lack of fluctuations in monitored values may be used as an indication that a sensor is damaged, and that the sensor may need to be remove or replaced.
[0164] In some embodiments, a sensor transmits (e.g., beacons) data to a receiver, e.g., a sensor or suite of sensors. The suite of sensors can also be referred to as an “ensemble of sensors.” The sensors in the suite of sensors can be analogous to those deployed in a space of an enclosure.
[0165] In some embodiments, a plurality of sensors is assembled into a sensor suite (e.g., sensor ensemble). At least two sensors of the plurality of sensors may be of a different type (e.g., are configured to measure different properties). Various sensor types can be assembled together (e.g., bundled up) and form a sensor suite. The plurality of sensors may be coupled to one electronic board. The electrical connection of at least two of the plurality of sensors in the sensor suit may be controlled (e.g., manually and/or automatically). For example, the sensor suite may be operatively coupled to, or comprise, a controller (e.g., a microcontroller). The controller may control and on/off connectivity of the sensor to electrical power. The controller can thus control the time (e.g., period) at which the sensor will be operative.
[0166] In some embodiments, one or more sensors are added or removed from a community of sensors, e.g., disposed in the enclosure and/or in the sensor suite (e.g., sensor ensemble). Newly added sensors may inform (e.g., beacon) other members of a community of sensor of its presence and relative location within a topology of the community. Examples of sensor community(ies) can be found, for example, in U.S. Provisional Patent Application Ser. No. 62/958,653, filed Jan. 8, 2020, titled “SENSOR AUTOLOCATION,” that is incorporated by reference herein in its entirety. Examples of sensor, and sensor ensembles can be found, for example, U.S. Provisional Patent Application Ser. No. 62/967,204, filed Jan. 29, 2020 titled “SENSOR CALLIBRATION AND OPERATION,” that is incorporated by reference herein in its entirety. These examples include methods of their use, software and apparatuses in which they are utilized and/or included.
[0167] Sensors of a sensor ensemble may be organized into a sensor module. A sensor ensemble may comprise a circuit board, such as a printed circuit board, in which a number of sensors are adhered or affixed to the circuit board. Sensors can be removed from a sensor module. For example, a sensor may be plugged and/or unplugged from the circuit board. Sensors may be individually activated and/or deactivated (e.g., using a switch). The circuit board may comprise a polymer. The circuit board may be transparent or non-transparent. The circuit board may comprise metal (e.g., elemental metal and/or metal alloy). The circuit board may comprise a conductor. The circuit board may comprise an insulator. The circuit board may comprise any geometric shape (e.g., rectangle or ellipse). The circuit board may be configured (e.g., may be of a shape) to allow the ensemble to be disposed in a mullion (e.g., of a window). The circuit board may be configured (e.g., may be of a shape) to allow the ensemble to be disposed in a frame (e.g., door frame and/or window frame). The mullion and/or frame may comprise one or more holes to allow the sensor(s) to obtain (e.g., accurate) readings. The circuit board may include an electrical connectivity port (e.g., socket). The circuit board may be connected to a power source (e.g., to electricity). The power source may comprise renewable or non-renewable power source.
[0168]
[0169] In some embodiments, an increase in the number and/or types of sensors may be used to increase a probability that one or more measured property is accurate and/or that a particular event measured by one or more sensor has occurred. In some embodiments, sensors of sensor ensemble may cooperate with one another. In an example, a radar sensor of sensor ensemble may determine presence of a number of individuals in an enclosure. A processor (e.g., processor 1315) may determine that detection of presence of a number of individuals in an enclosure is positively correlated with an increase in carbon dioxide concentration. In an example, the processor-accessible memory may determine that an increase in detected infrared energy is positively correlated with an increase in temperature as detected by a temperature sensor. In some embodiments, network interface (e.g., 1350) may communicate with other sensor ensembles similar to sensor ensemble. The network interface may additionally communicate with a controller.
[0170] Individual sensors (e.g., sensor 1310A, sensor 1310D, etc.) of a sensor ensemble may comprise and/or utilize at least one dedicated processor. A sensor ensemble may utilize a remote processor (e.g., 1354) utilizing a wireless and/or wired communications link. A sensor ensemble may utilize at least one processor (e.g., processor 1352), which may represent a cloud-based processor coupled to a sensor ensemble via the cloud (e.g., 1350). Processors (e.g., 1352 and/or 1354) may be located in the same building, in a different building, in a building owned by the same or different entity, a facility owned by the manufacturer of the window/controller/sensor ensemble, or at any other location. In various embodiments, as indicated by the dotted lines of
[0171] In some embodiments, a plurality of sensors of the same type may be distributed in an enclosure. At least one of the plurality of sensors of the same type, may be part of an ensemble. For example, at least two of the plurality of sensors of the same type, may be part of at least two ensembles. The sensor ensembles may be distributed in an enclosure. An enclosure may comprise a conference room. For example, a plurality of sensors of the same type may measure an environmental parameter in the conference room. Responsive to measurement of the environmental parameter of an enclosure, a parameter topology of the enclosure may be generated. A parameter topology may be generated utilizing output signals from any type of sensor of sensor ensemble, e.g., as disclosed herein. Parameter topologies may be generated for any enclosure of a facility such as conference rooms, hallways, bathrooms, cafeterias, garages, auditoriums, utility rooms, storage facilities, equipment rooms, and/or elevators.
[0172]
[0173] In particular embodiments, one or more sensors of the sensor ensemble provide readings. In some embodiments, the sensor is configured to sense and/or identify a parameter. The parameter may comprise temperature, particulate matter, volatile organic compounds, electromagnetic energy, pressure, acceleration, time, radar, lidar, glass breakage, movement, or gas. The gas may comprise a Nobel gas. The gas may be a gas harmful to an average human. The gas may be a gas present in the ambient atmosphere (e.g., oxygen, carbon dioxide, ozone, chlorinated carbon compounds, or nitrogen). The gas may comprise radon, carbon monoxide, hydrogen sulfide, hydrogen, oxygen, water (e.g., humidity). The electromagnetic sensor may comprise an infrared, visible light, ultraviolet sensor. The infrared radiation may be passive infrared radiation (e.g., black body radiation). The electromagnetic sensor may sense radio waves. The radio waves may comprise wide band, or ultra-wideband radio signals. The radio waves may comprise pulse radio waves. The radio waves may comprise radio waves utilized in communication. The gas sensor may sense a gas type, flow (e.g., velocity and/or acceleration), pressure, and/or concentration. The readings may have an amplitude range. The readings may have a parameter range. For example, the parameter may be electromagnetic wavelength, and the range may be a range of detected wavelengths.
[0174] In some embodiments, the sensor data is responsive to the environment in the enclosure and/or to any inducer(s) of a change (e.g., any environmental disruptor) in this environment. The sensors data may be responsive to emitters operatively coupled to (e.g., in) the enclosure (e.g., an occupant, appliances (e.g., heater, cooler, ventilation, and/or vacuum), opening). For example, the sensor data may be responsive to an air conditioning duct, or to an open window. The sensor data may be responsive to an activity taking place in the room. The activity may include human activity, and/or non-human activity. The activity may include electronic activity, gaseous activity, and/or chemical activity. The activity may include a sensual activity (e.g., visual, tactile, olfactory, auditory, and/or gustatory). The activity may include an electronic and/or magnetic activity. The activity may be sensed by a person. The activity may not be sensed by a person. The sensors data may be responsive to the occupants in the enclosure, substance (e.g., gas) flow, substance (e.g., gas) pressure, and/or temperature.
[0175] In one example, sensor ensembles 1405A, 1405B, and 1405C include carbon dioxide (CO.sub.2) sensor, and an ambient noise sensor. A carbon dioxide sensor of sensor ensemble 1405A may provide a reading as depicted in sensor output reading profile 1425A. A noise sensor of sensor ensemble 1405A may provide a reading also depicted in sensor output reading profile 1425A. A carbon dioxide sensor of sensor ensemble 1405B may provide a reading as depicted in sensor output reading profile 1425B. A noise sensor of sensor ensemble 1405B may provide a reading also as depicted in sensor output reading profile 1425B. Sensor output reading profile 1425B may indicate higher levels of carbon dioxide and noise relative to sensor output reading profile 1425A. Sensor output reading profile 1425C may indicate lower levels of carbon dioxide and noise relative to sensor output reading profile 1425B. Sensor output reading profile 1425C may indicate carbon dioxide and noise levels similar to those of sensor output reading profile 1425A. Sensor output reading profiles 1425A, 1425B, and 1425C may comprise indications representing other sensor readings, such as temperature, humidity, particulate matter, volatile organic compounds, ambient light, pressure, acceleration, time, radar, lidar, ultra-wideband radio signals, passive infrared, and/or glass breakage, movement detectors.
[0176] In some embodiments, data from a sensor in a sensor in the enclosure (e.g., and in the sensor ensemble) is collected and/or processed (e.g., analyzed). The data processing can be performed by a processor of the sensor, by a processor of the sensor ensemble, by another sensor, by another ensemble, in the cloud, by a processor of the controller, by a processor in the enclosure, by a processor outside of the enclosure, by a remote processor (e.g., in a different facility), by a manufacturer (e.g., of the sensor, of the window, and/or of the building network). The data of the sensor may have a time indicator (e.g., may be time stamped). The data of the sensor may have a sensor location identification (e.g., be location stamped). The sensor may be identifiably coupled with one or more controllers.
[0177] In particular embodiments, sensor output reading profiles 1425A, 1425B, and 1425C may be processed. For example, as part of the processing (e.g., analysis), the sensor output reading profiles may be plotted on a graph depicting a sensor reading as a function of a dimension (e.g., the “X” dimension) of an enclosure (e.g., conference room 1402). In an example, a carbon dioxide level indicated in sensor output reading profile 1425A may be indicated as point 1435A of CO.sub.2 graph 1430 of
[0178] In some embodiments, processing data derived from the sensor comprises applying one or more models. The models may comprise mathematical models. The processing may comprise fitting of models (e.g., curve fitting). The model may be multi-dimensional (e.g., two or three dimensional). The model may be represented as a graph (e.g., 2 or 3 dimensional graph). For example, the model may be represented as a contour map (e.g., as depicted in
[0179] In particular embodiments, sensor ensembles 1405A, 1405B, and/or 1405C, may be capable of accessing a model to permit curve fitting of sensor readings as a function of one or more dimensions of an enclosure. In an example, a model may be accessed to generate sensor profile curves 1450A, 1450B, 1450C, 1450D, and 1450E, utilizing points 1435A, 1435B, and 1435C of CO.sub.2 graph 1430. In an example, a model may be accessed to generate sensor profile curves 1451A, 1451B, 1451C, 1451B, and 1451E utilizing points 1445A, 1445B, and 1445C of noise graph 1440. Additional models may utilize additional readings from sensor ensembles (e.g., 1405A, 1405B, and/or 1405C) to provide curves in addition to sensor profile curves 1450 and 1451 of
[0180] In certain embodiments, one or more models utilized to form curves 1450A-1450E and 1451A-1451E) may provide a parameter topology of an enclosure. In an example, a parameter topology (as represented by curves 1450A-1450E and 1451A-1451E) may be synthesized or generated from sensor output reading profiles. The parameter topology may be a topology of any sensed parameter disclosed herein. In an example, a parameter topology for a conference room (e.g., conference room 1402) may comprise a carbon dioxide profile having relatively low values at locations away from a conference room table and relatively high values at locations above (e.g., directly above) a conference room table. In an example, a parameter topology for a conference room may comprise a multi-dimensional noise profile having relatively low values at locations away from a conference table and slightly higher values above (e.g., directly above) a conference room table.
[0181] In some embodiments, sensor data corresponding to an individual is collected upon arrival and an entry station or admittance location to an enclosure (e.g., facility). See for example,
[0182] In some embodiments, sensor data is collected, compiled, and related to the individual person for whom measurements are obtained. The relation may comprise identification (ID) of the individual. The identification may or may not comprise personal data of the individual (e.g., name, home address, phone number, governmental identification number, fingerprint, retina scan, bodily features, and/or facial features). Sensor data (e.g., comprising raw and/or processed measurements) may be communicated to a database (e.g., via a controller or compiler) along with a respective sensor ID, and a time stamp (e.g., including time, and/or date). The person can be identified with a tag comprising geo-location technology, image processing of a person's face, size, gait, blood pressure, infrared (IR) signature, heart rate and/or any other personal characteristics (e.g., using radar, IR and the like). The geo-location technology may comprise radio frequency (e.g., RFID) chip. The geo-location technology may comprise BLE, GPS, or UWB technology. The communication can be to at least one software application (e.g., executing in a mobile device of the individual or of another person) and/or to a control system operatively coupled to the enclosure (e.g., building). The application (e.g., control system and/or phone app) may keep a table (for one or more users) of sensor IDs, bodily characteristic (e.g., temperature), and time stamp (e.g., time and date). Data may be collected overtime to determine a “normal” for each uniquely identified individual (e.g., building occupant). The collected data can be used as a baseline for deviation from the norm of that individual. Analysis of the collected data in the database relating to an individual may be performed on the basis of relative sensor data (e.g., for a particular sensor). For example, relative sensor measurements of the person collected at different times may be compared. Since relative measurements are being used, the sensor from which the data is obtained may not need to be calibrated. Absolute values of the sensor data may be immaterial as long as the increments measured by the sensor are accurate (e.g., as long as the relative measurements are accurate).
[0183] In some embodiments, if a relative difference between sensor measurement exceeds a threshold, an event is triggered (e.g., a notification event). In order to ascertain a relative difference threshold to be used as a trigger that identifies an abnormal condition, a person can be monitored overtime to learn personal bodily characteristic behavior (e.g., oscillations) and behavioral patterns. Monitoring may be performed using a learning module. The learning module may include artificial intelligence (Ai) comprised of machine learning. In one example, an individual occupant and/or user of an enclosure (e.g., wherein the user has a unique identifier such as “123”) may routinely work out at lunch time, and thus their temperature and heartrate may usually increase dramatically during that time period. The Ai may take this into account when considering an adverse variation in characteristics for user #123. For example, a temperature, blood pressure, heart rate, etc., of the individual may change according to normal cycles during the course of a day; and these patterns (which may be unique for the individual) may be recorded and norms for the corresponding individual may be established. Seasonal variables and/or other extrinsic factors (referred to herein as “paradigms”) can be input to the ML for improved accuracy. Some such paradigms can be quantified when sensor data is collected. Quantified values can be stored together with the sensor data for use in data analysis, e.g., to determine normal ranges of bodily characteristics of the individual during various paradigms. Once an abnormal condition has been detected for an individual, a report may be generated. Optionally, a notification system may be activated to provide a notification to the affected individual. In some embodiments, notifications are (e.g., also) sent to contactees who may have been exposed to the affected individual.
[0184]
[0185] A controller system includes modules for ID and location tracking 1503 and a data compiler 1504. Using sensor data from sensors 1502 and user identifiers from ID tracking module 1503, data compiler 1504 organizes the sensor data according to individuals (e.g., users), sensor ID, time, and date, to support the analysis of a bodily characteristic of each tracked user over a span of time. The organized data can be stored in a personal database 1505 and/or a collective database 1506. Personal database 1505 may be stored in a mobile device carried by the user (e.g., a smartphone executing the corresponding app) or other personal device such as a laptop computer. Collective database 1506 may be stored in a networked controller in the facility or external to the facility (e.g., remotely via a cloud). Use of collective database 1506 may be beneficial in connection with tracking multiple users, and optionally performing notifications and contact tracing as described below (e.g.,
[0186] For users represented in database 1505 and/or 1506, the stored sensor data is analyzed according to relative changes of bodily characteristics 1507. The analysis can use learning module (e.g., artificial intelligence comprising machine learning). Learning module 1507 may be implemented in the networked controller or in the dedicated user device (e.g., smartphone). In some embodiments, when new sensor data for a corresponding user is obtained and entered into database 1505 or 1506, the analysis module 1507 makes a determination whether sufficient data is stored in order to identify a relevant normal for the sensed characteristic that can be used for a comparison. Seasonal data 1508, other environmental, and/or other circumstantial factors may be provided to analysis module 1507, e.g., to improve determinations of the appropriate normal to be applied. A report is provided in 1509 by the analysis module. When the newly collected sensor data indicates an abnormal condition (e.g., a difference between the new sensor data and the calculated norm is greater than a threshold), a notification system 1510 is optionally activated, e.g., to send out various notifications (e.g., a text message or email) to an affected user 1511, and/or contactees 1512 who may have been in proximity to user 1511 during the abnormal condition (e.g., for a time period above a time-threshold), and/or to a central administrator and/or health official 1513.
[0187]
[0188] In some embodiments, tracking of users in a facility provides a basis for identifying potential contactees of users for whom abnormal bodily characteristics are detected. For example, an opt-in building-secure contact tracing system may store times and locations of user movements for correlation. A tracing system may utilize (1) occupancy data for (e.g., anonymized or identifiable) application users, and (2) adequate data infrastructure for retaining the timeframes of their presence. As part of, or separately from, the user tracking and compilation of sensor data for users during their occupancy in a facility (e.g., including at least one building), each user may carry a user device that automatically connects to a network infrastructure upon entering the facility. Based at least in part on interaction with the user device(s), the timeframes of user occupancy can be locally and/or remotely retained (e.g., stored in a tracking database). In some embodiments, tracking of users is based at least in part on communication between the user carried ID device (e.g., smartphone, RFID badge, or laptop) and a wireless and/or wired network infrastructure based at least in part on geo-localization methods such as GPS tracking, ranging, triangulation, WiFi presence, short-range communication such as UWB, and/or other methods. In some embodiments, user tracking is accomplished without a user carried device, e.g., by utilizing remote sensing and identification relying on sensors (e.g., in device ensembles) deployed throughout a facility (e.g. using as facial recognition).
[0189] In some embodiments, anonymized or non-anonymized tracking statistics are maintained over a rolling time period (e.g., several days, a week, several weeks, months, or years). For each person with a registered user ID, locations and times in the facility may be stored. When a particular user is identified with an abnormal condition, the locations and times of their movements in the building may be retrieved as search criteria for extracting the user IDs and times for other facility occupants which converge with the affected user (e.g., within a distance and optionally for a time above a time-threshold). A push notification to other facility users (e.g., coworkers or cohabitants) on site in a timeframe of potential exposure, may occur in an anonymized or identified fashion. In some embodiments, local retention of potential exposure timeframes serves to facilitate adherence to medical guidelines (e.g., in the jurisdiction) on proper quarantining and/or sanitation procedures, e.g., enabling a targeted response to affected person(s) and/or to those facility areas most frequented by the affected person(s).
[0190]
[0191] In some embodiments, the bodily characteristic tracking sensors may be disposed throughout a facility. There may be a plurality of sensor types required to (e.g., accurately) track one characteristics of an individual throughout the facility. There may be a single sensor type required to (e.g., accurately) track one characteristics of an individual throughout the facility. In some embodiments, the sensor network tracks one bodily characteristic of individual(s) in the facility. In some embodiments, the sensor network tracks a plurality of different bodily characteristics of individual(s) in the facility. The tracking a bodily characteristic of an individual in the facility may comprise recording an identification of an individual, location of the individual in the facility, time and date of the measurement, a type of sensor measurement (e.g., infrared sensor measurements), and optionally another type of sensor measurement (if required to accurately reflect the bodily characteristic of the individual, e.g., visible sensor measurements).
[0192]
[0193] In some embodiments, the availability of tracking data for facility occupants is utilized to aid in maintenance of physical distances between occupants (e.g., maintain social distancing above a distance threshold). A user may notify the network (e.g., using a software application—app) of his/her destination. For example, a user sitting at an office desk would like to go to conference room X. The app may use the tracking data and suggest an optimal route that is the one least crowded with occupant, from the current location of the occupant (e.g., desk) to the occupant destination in the facility (e.g., conference room X). The app may use the projected analysis (e.g., using machine learning, occupancy schedule for the facility, and/or activity schedule for the facility) to foresee occupancy of the enclosure during the requested movement time of the requesting user. The app may suggest an optimal route that is the one least crowded with occupant during the expected travel time, from the current location of the occupant (e.g., desk) to the occupant destination in the facility (e.g., conference room X). When the user begins travel from their location (e.g., desk) toward their requested destination (e.g., conference room X), the controller may review a tracking database to check for clusters of other occupants along the projected path (e.g., in real time, during the occupant's travel). When such a cluster is found, alternative routes may be evaluated to find and suggest another available route to the destinations, which is less crowded. Such less crowded route option may be automatically shared with the user by the app (e.g., pushed to the user).
[0194]
[0195] In some embodiments, the availability of tracking data for facility occupants is utilized to aid in maintenance of physical distances between occupants (e.g., social distancing). As occupants are tracked within a facility, a learning module may compile the typical movements of individuals for various times of day and/or the occurrences of repeating events. Using the learned tendencies for movements of users, the learning module can anticipate where people are going as they move about. When a projected path of movement for a particular user is discerned by the learning module, a controller (e.g., using a processor) can evaluate conditions (e.g., crowded or not) on the expected path. In the event that conditions are crowded in the anticipated path of the user, a (e.g., push) notification may be sent to the user with a recommended route to avoid crowded areas (e.g., to maintain social distancing of the user during his/her taking the path). For example, a user may have a tracking history showing a recurring pattern of moving from their desk to a printer station, from the printer station to a file room, and then from the file room back to their desk. When the user begins travel from their desk toward the printer station, the controller may anticipate such a round trip and then review a tracking database to check for clusters of other occupants along the projected path. When such a cluster is found, alternative routes may be evaluated to find another available route to the destinations which is less or least crowded (e.g., to avoid other people). The alternate route can be automatically shared with the user by the app.
[0196]
[0197] In some embodiments, environmental characteristics are monitored using any of the described sensor modalities to find abnormal characteristics of people and/or surfaces in a facility. For example, humidity changes may be monitored as an indication of excessive perspiration of a sick person. Similarity, gases or chemical concentrations can be monitored as indication an of various kinds of illness. In some embodiments, the sensors or sensor ensembles deployed in a facility are used to sense and/or identify the temperatures of designated surfaces that are targets for regular disinfection (e.g., furniture and/or fixture surfaces such as countertops, doors, table-tops, handles, windows, frames, etc.). Such surfaces may be chosen as a result of being significant reservoirs for infectious agents (e.g., pathogens) to collect and potentially be passed on to other occupants, e.g., if not disinfected. The surfaces can be targeted for regular cleaning and monitoring. Monitoring of the surface can be initiated manually or by otherwise detecting a cleaning activity. After cleaning (e.g., using typical disinfectants and/or other liquid) evaporation of the cleaning liquids (e.g. solvents) from the surface may cool the surface. A cleaning event may be detected by measuring surface temperature at a plurality of sample times and comparing a plurality of consecutive samples (e.g., two or more) to detect a temperature drop (e.g., above a threshold and/or temperature drop rate), for example, as the temperature of the surface gets reduced due to evaporation. Once a cleaning event has been identified, an elapsed time since the last cleaning of the surface can be detected. In some embodiments, a surface temperature may be intermittently or (e.g., substantially) continuously measured (e.g., at a predetermined sample rate) to determine how much time elapsed since the last cleaning, and optionally whether the surface requires a second cleaning (e.g., as too much time elapsed from the last cleaning event).
[0198]
[0199] In some embodiments, the sensor(s) are operatively coupled to at least one controller and/or processor. Sensor readings may be obtained by one or more processors and/or controllers. A controller may comprise a processing unit (e.g., CPU or GPU). A controller may receive an input (e.g., from at least one sensor). The controller may comprise circuitry, electrical wiring, optical wiring, socket, and/or outlet. A controller may deliver an output. A controller may comprise multiple (e.g., sub-) controllers. The controller may be a part of a control system. A control system may comprise a master controller, floor (e.g., comprising network controller) controller, a local controller. The local controller may be a window controller (e.g., controlling an optically switchable window), enclosure controller, or component controller. For example, a controller may be a part of a hierarchal control system (e.g., comprising a main controller that directs one or more controllers, e.g., floor controllers, local controllers (e.g., window controllers), enclosure controllers, and/or component controllers). A physical location of the controller type in the hierarchal control system may be changing. For example: At a first time: a first processor may assume a role of a main controller, a second processor may assume a role of a floor controller, and a third processor may assume the role of a local controller. At a second time: the second processor may assume a role of a main controller, the first processor may assume a role of a floor controller, and the third processor may remain with the role of a local controller. At a third time: the third processor may assume a role of a main controller, the second processor may assume a role of a floor controller, and the first processor may assume the role of a local controller. A controller may control one or more devices (e.g., be directly coupled to the devices). A controller may be disposed proximal to the one or more devices it is controlling. For example, a controller may control an optically switchable device (e.g., IGU), an antenna, a sensor, and/or an output device (e.g., a light source, sounds source, smell source, gas source, HVAC outlet, or heater). In one embodiment, a floor controller may direct one or more window controllers, one or more enclosure controllers, one or more component controllers, or any combination thereof. The floor controller may comprise a floor controller. For example, the floor (e.g., comprising network) controller may control a plurality of local (e.g., comprising window) controllers. A plurality of local controllers may be disposed in a portion of a facility (e.g., in a portion of a building). The portion of the facility may be a floor of a facility. For example, a floor controller may be assigned to a floor. In some embodiments, a floor may comprise a plurality of floor controllers, e.g., depending on the floor size and/or the number of local controllers coupled to the floor controller. For example, a floor controller may be assigned to a portion of a floor. For example, a floor controller may be assigned to a portion of the local controllers disposed in the facility. For example, a floor controller may be assigned to a portion of the floors of a facility. A master controller may be coupled to one or more floor controllers. The floor controller may be disposed in the facility. The master controller may be disposed in the facility, or external to the facility. The master controller may be disposed in the cloud. A controller may be a part of, or be operatively coupled to, a building management system. A controller may receive one or more inputs. A controller may generate one or more outputs. The controller may be a single input single output controller (SISO) or a multiple input multiple output controller (MIMO). A controller may interpret an input signal received. A controller may acquire data from the one or more components (e.g., sensors). Acquire may comprise receive or extract. The data may comprise measurement, estimation, determination, generation, or any combination thereof. A controller may comprise feedback control. A controller may comprise feed-forward control. Control may comprise on-off control, proportional control, proportional-integral (PI) control, or proportional-integral-derivative (PID) control. Control may comprise open loop control, or closed loop control. A controller may comprise closed loop control. A controller may comprise open loop control. A controller may comprise a user interface. A user interface may comprise (or operatively coupled to) a keyboard, keypad, mouse, touch screen, microphone, speech recognition package, camera, imaging system, or any combination thereof. Outputs may include a display (e.g., screen), speaker, or printer.
[0200]
[0201] In particular embodiments, sensor readings from a particular sensor may be correlated with sensor readings from a sensor of the same type or of a different type. Receipt of a sensor reading may give rise to a sensor accessing correlation data from other sensors disposed within the same enclosure. Based, at least in part, on the access correlation data, the reliability of a sensor may be determined or estimated. Responsive to determination or estimation of sensor reliability, a sensor output reading may be adjusted (e.g., increased/decreased). A reliability value may be assigned to a sensor based on adjusted sensor readings.
[0202] A sensor reading may be any type of reading, such as detection of movement of individuals within an enclosure, temperature, humidity, or any other property detected by the sensor. Sensor readings may be correlated with correlation data. The correlation data may be accessed from other sensors disposed in the enclosure. Correlation data may relate to output readings from a sensor of the same type or a sensor of a different type operating within the enclosure. In an example, a noise sensor may access data from a movement sensor to determine if one or more individuals have entered an enclosure. One or more individuals moving within an enclosure may emit a level of noise. In an example, output signals from a noise sensor may be corroborated by a second noise sensor and/or by a movement detector. Based at least in part on the accessed correlation data, ran analysis (e.g., evaluation) of the environmental characteristic at one or more position of the environment may be made. The one or more positions may correlate to a position of an individual in the environment (e.g., to detect any abnormal bodily characteristic of the individual(s). The environmental characteristic(s) perturbed by the individual(s) in the environment may be detected for any abnormal bodily characteristic. Once the environmental characteristic(s) detected are analyzed for any abnormal bodily characteristic, they may be reported (e.g., as disclosed herein).
[0203]
[0204] In some embodiments, the sensor data may be time dependent. In some embodiments, the sensor data may be space dependent. The model may utilize time and/or space dependency of the sensed parameter. A model generator may permit fitting of sensor readings as a function of one or more dimensions of an enclosure. In an example, a model provides sensor profile curves for carbon dioxide may utilize various gaseous diffusion models, which may allow prediction of a level of carbon dioxide at points in between sensor locations. Processor and memory may facilitate processing of models.
[0205] In some embodiments, the sensor and/or sensor ensemble may act as an event detector. The event detector may operate to direct activity of sensors in an enclosure. In an example, in response to event detector determining that very few individuals remain in an enclosure, event detector may direct carbon dioxide sensors to reduce a sampling rate. Reduction of a sampling rate may extend the life of a sensor (e.g., a carbon dioxide sensor). In another example, in response to event detector determining that a large number of individuals are present in a room, event detector may increase the sampling rate of a carbon dioxide sensor. In an example, in response to event detector receiving a signal from a glass breakage sensor, event detector may activate one or more movement detectors of an enclosure, one or more radar units of a detector. A network interface (e.g., 2350) may be configured or designed to communicate with one or more sensors via wireless communications links, wired communications links, or any combination thereof.
[0206] The controller may monitor and/or direct (e.g., physical) alteration of the operating conditions of the apparatuses, software, and/or methods described herein. Control may comprise regulate, manipulate, restrict, direct, monitor, adjust, modulate, vary, after, restrain, check, guide, or manage. Controlled (e.g., by a controller) may include attenuated, modulated, varied, managed, curbed, disciplined, regulated, restrained, supervised, manipulated, and/or guided. The control may comprise controlling a control variable (e.g. temperature, power, voltage, and/or profile). The control can comprise real time or off-line control. A calculation utilized by the controller can be done in real time, and/or offline. The controller may be a manual or a non-manual controller. The controller may be an automatic controller. The controller may operate upon request. The controller may be a programmable controller. The controller may be programed. The controller may comprise a processing unit (e.g., CPU or GPU). The controller may receive an input (e.g., from at least one sensor). The controller may deliver an output. The controller may comprise multiple (e.g., sub-) controllers. The controller may be a part of a control system. The control system may comprise a master controller, floor controller, local controller (e.g., enclosure controller, or window controller). The controller may receive one or more inputs. The controller may generate one or more outputs. The controller may be a single input single output controller (SISO) or a multiple input multiple output controller (MIMO). The controller may interpret the input signal received. The controller may acquire data from the one or more sensors. Acquire may comprise receive or extract. The data may comprise measurement, estimation, determination, generation, or any combination thereof. The controller may comprise feedback control. The controller may comprise feed-forward control. The control may comprise on-off control, proportional control, proportional-integral (PI) control, or proportional-integral-derivative (PID) control. The control may comprise open loop control, or closed loop control. The controller may comprise closed loop control. The controller may comprise open loop control. The controller may comprise a user interface. The user interface may comprise (or operatively coupled to) a keyboard, keypad, mouse, touch screen, microphone, speech recognition package, camera, imaging system, or any combination thereof. The outputs may include a display (e.g., screen), speaker, or printer.
[0207] The methods, systems and/or the apparatus described herein may comprise a control system. The control system can be in communication with any of the apparatuses (e.g., sensors) described herein. The sensors may be of the same type or of different types, e.g., as described herein. For example, the control system may be in communication with the first sensor and/or with the second sensor. The control system may control the one or more sensors. The control system may control one or more components of a building management system (e.g., lightening, security, and/or air conditioning system). The controller may regulate at least one (e.g., environmental) characteristic of the enclosure. The control system may regulate the enclosure environment using any component of the building management system. For example, the control system may regulate the energy supplied by a heating element and/or by a cooling element. For example, the control system may regulate velocity of an air flowing through a vent to and/or from the enclosure. The control system may comprise a processor. The processor may be a processing unit. The controller may comprise a processing unit. The processing unit may be central. The processing unit may comprise a central processing unit (abbreviated herein as “CPU”). The processing unit may be a graphic processing unit (abbreviated herein as “GPU”). The controller(s) or control mechanisms (e.g., comprising a computer system) may be programmed to implement one or more methods of the disclosure. The processor may be programmed to implement methods of the disclosure. The controller may control at least one component of the forming systems and/or apparatuses disclosed herein.
[0208]
[0209] The computer system can include a processing unit (e.g., 2406) (also “processor,” “computer and “computer processor” used herein). The computer system may include memory or memory location (e.g., 2402) (e.g., random-access memory, read-only memory, flash memory), electronic storage unit (e.g., 2404) (e.g., hard disk), communication interface (e.g., 2403) (e.g., network adapter) for communicating with one or more other systems, and peripheral devices (e.g., 2405), such as cache, other memory, data storage and/or electronic display adapters. In the example shown in
[0210] The processing unit can execute a sequence of machine-readable instructions, which can be embodied in a program or software. The instructions may be stored in a memory location, such as the memory 2402. The instructions can be directed to the processing unit, which can subsequently program or otherwise configure the processing unit to implement methods of the present disclosure. Examples of operations performed by the processing unit can include fetch, decode, execute, and write back. The processing unit may interpret and/or execute instructions. The processor may include a microprocessor, a data processor, a central processing unit (CPU), a graphical processing unit (GPU), a system-on-chip (SOC), a co-processor, a network processor, an application specific integrated circuit (ASIC), an application specific instruction-set processor (ASIPs), a controller, a programmable logic device (PLD), a chipset, a field programmable gate array (FPGA), or any combination thereof. The processing unit can be part of a circuit, such as an integrated circuit. One or more other components of the system 1800 can be included in the circuit.
[0211] The storage unit can store files, such as drivers, libraries and saved programs. The storage unit can store user data (e.g., user preferences and user programs). In some cases, the computer system can include one or more additional data storage units that are external to the computer system, such as located on a remote server that is in communication with the computer system through an intranet or the Internet.
[0212] The computer system can communicate with one or more remote computer systems through a network. For instance, the computer system can communicate with a remote computer system of a user (e.g., operator). Examples of remote computer systems include personal computers (e.g., portable PC), slate or tablet PC's (e.g., Apple® iPad, Samsung® Galaxy Tab), telephones, Smart phones (e.g., Apple® iPhone, Android-enabled device, Blackberry®), or personal digital assistants. A user (e.g., client) can access the computer system via the network.
[0213] Methods as described herein can be implemented by way of machine (e.g., computer processor) executable code stored on an electronic storage location of the computer system, such as, for example, on the memory 2402 or electronic storage unit 2404. The machine executable or machine-readable code can be provided in the form of software. During use, the processor 2406 can execute the code. In some cases, the code can be retrieved from the storage unit and stored on the memory for ready access by the processor. In some situations, the electronic storage unit can be precluded, and machine-executable instructions are stored on memory.
[0214] The code can be pre-compiled and configured for use with a machine have a processer adapted to execute the code or can be compiled during runtime. The code can be supplied in a programming language that can be selected to enable the code to execute in a pre-compiled or as-compiled fashion.
[0215] In some embodiments, the processor comprises a code. The code can be program instructions. The program instructions may cause the at least one processor (e.g., computer) to direct a feed forward and/or feedback control loop. In some embodiments, the program instructions cause the at least one processor to direct a closed loop and/or open loop control scheme. The control may be based at least in part on one or more sensor readings (e.g., sensor data). One controller may direct a plurality of operations. At least two operations may be directed by different controllers. In some embodiments, a different controller may direct at least two of operations (a), (b) and (c). In some embodiments, different controllers may direct at least two of operations (a), (b) and (c). In some embodiments, a non-transitory computer-readable medium cause each a different computer to direct at least two of operations (a), (b) and (c). In some embodiments, different non-transitory computer-readable mediums cause each a different computer to direct at least two of operations (a), (b) and (c). The controller and/or computer readable media may direct any of the apparatuses or components thereof disclosed herein. The controller and/or computer readable media may direct any operations of the methods disclosed herein.
[0216] In some embodiments, the at least one sensor is operatively coupled to a control system (e.g., computer control system). The sensor may comprise light sensor, acoustic sensor, vibration sensor, chemical sensor, electrical sensor, magnetic sensor, fluidity sensor, movement sensor, speed sensor, position sensor, pressure sensor, force sensor, density sensor, distance sensor, or proximity sensor. The sensor may include temperature sensor, weight sensor, material (e.g., powder) level sensor, metrology sensor, gas sensor, or humidity sensor. The metrology sensor may comprise measurement sensor (e.g., height, length, width, angle, and/or volume). The metrology sensor may comprise a magnetic, acceleration, orientation, or optical sensor. The sensor may transmit and/or receive sound (e.g., echo), magnetic, electronic, or electromagnetic signal. The electromagnetic signal may comprise a visible, infrared, ultraviolet, ultrasound, radio wave, or microwave signal. The gas sensor may sense any of the gas delineated herein. The distance sensor can be a type of metrology sensor. The distance sensor may comprise an optical sensor, or capacitance sensor. The temperature sensor can comprise Bolometer, Bimetallic strip, calorimeter, Exhaust gas temperature gauge, Flame detection, Gardon gauge, Golay cell, Heat flux sensor, Infrared thermometer, Microbolometer, Microwave radiometer, Net radiometer, Quartz thermometer, Resistance temperature detector, Resistance thermometer, Silicon band gap temperature sensor, Special sensor microwave/imager, Temperature gauge, Thermistor, Thermocouple, Thermometer (e.g., resistance thermometer), or Pyrometer. The temperature sensor may comprise an optical sensor. The temperature sensor may comprise image processing. The temperature sensor may comprise a camera (e.g., IR camera, CCD camera). The pressure sensor may comprise Barograph, Barometer, Boost gauge, Bourdon gauge, Hot filament ionization gauge, Ionization gauge, McLeod gauge, Oscillating U-tube, Permanent Downhole Gauge, Piezometer, Pirani gauge, Pressure sensor, Pressure gauge, Tactile sensor, or Time pressure gauge. The position sensor may comprise Auxanometer, Capacitive displacement sensor, Capacitive sensing, Free fall sensor, Gravimeter, Gyroscopic sensor, Impact sensor, Inclinometer, Integrated circuit piezoelectric sensor, Laser rangefinder, Laser surface velocimeter, LIDAR, Linear encoder, Linear variable differential transformer (LVDT), Liquid capacitive inclinometers, Odometer, Photoelectric sensor, Piezoelectric accelerometer, Rate sensor, Rotary encoder, Rotary variable differential transformer, Selsyn, Shock detector, Shock data logger, Tilt sensor, Tachometer, Ultrasonic thickness gauge, Variable reluctance sensor, or Velocity receiver. The optical sensor may comprise a Charge-coupled device, Colorimeter, Contact image sensor, Electro-optical sensor, Infra-red sensor, Kinetic inductance detector, light emitting diode (e.g., light sensor), Light-addressable potentiometric sensor, Nichols radiometer, Fiber optic sensor, Optical position sensor, Photo detector, Photodiode, Photomultiplier tubes, Phototransistor, Photoelectric sensor, Photoionization detector, Photomultiplier, Photo resistor, Photo switch, Phototube, Scintillometer, Shack-Hartmann, Single-photon avalanche diode, Superconducting nanowire single-photon detector, Transition edge sensor, Visible light photon counter, or Wave front sensor. The one or more sensors may be connected to a control system (e.g., to a processor, to a computer). The sensor may comprise complementary metal oxide semiconductor (CMOS).
[0217] In various embodiments, a network infrastructure supports a control system for one or more windows such as electrochromic (e.g., tintable) windows. The control system may comprise one or more controllers operatively coupled (e.g., directly or indirectly) to one or more windows. While the disclosed embodiments describe electrochromic windows (also referred to herein as “optically switchable windows,” “tintable windows”, or “smart windows”), the concepts disclosed herein may apply to other types of switchable optical devices including, for example, a liquid crystal device, or a suspended particle device. For example, a liquid crystal device and/or a suspended particle device may be implemented instead of, or in addition to, an electrochromic device. Tintable windows may operate using liquid crystal devices, suspended particle devices, microelectromechanical systems (MEMS) devices (such as microshutters), or any technology known now, or later developed, that is configured to control light transmission through a window. Windows (e.g., with MEMS devices for tinting) are described in U.S. patent application Ser. No. 14/443,353, filed May 15, 2015, titled “MULTI-PANE WINDOWS INCLUDING ELECTROCHROMIC DEVICES AND ELECTROMECHANICAL SYSTEMS DEVICES,” that is incorporated herein by reference in its entirety. In some cases, one or more tintable windows can be located within the interior of a building, e.g., between a conference room and a hallway. In some cases, one or more tintable windows can be used in automobiles, trains, aircraft, and other vehicles, e.g., in lieu of a passive and/or non-tinting window.
[0218] In some embodiments, bodily characteristics of an individual may be performed at certain locations of the body. For example, temperature may be measured on the forehead. At least one sensor may be disposed in a structure (e.g., fixture) to measure the bodily characteristic. The sensor may focus on the bodily location to sense and/or identify the bodily characteristic. For example, the sensor may focus on the forehead to measure the temperature. The sensor may be directed to focus on a certain lateral (e.g., horizontal) distance from the sensor. The distance may correspond to the distance of the bodily location. The distance may correspond to a designated location of the individual from the sensor. The sensor may be directed to focus on a certain vertical distance from the sensor. The vertical distance may correspond to the vertical distance of the bodily location. The distance may correspond to an average location of the bodily location in the population, e.g., depending on age group and/or gender.
[0219] In some embodiments, the sensor(s) are operatively coupled to a network. At least one controller operatively coupled to network (e.g., one or more processors operatively coupled to the network) may direct the sensor(s) to focus on the bodily location based at least in part on image recognition of facial landmarks. Facial landmark recognition may utilized one or more (e.g., other) sensors such as sensors for visible and/or IR sensors. The sensors may be configured to identify the individual based at least in part on (i) the individual's temperature as compared to the surrounding and/or background, (ii) facial landmark features of the individual. The lateral (e.g., horizontal) distance of the individual from the sensor(s) may be estimated using distance between facial features and/or size of facial features, and their comparison to an average population. For example, distance between eyes. For example, size of a pupil. The lateral distance may be estimated using a combination of IR sensor and visible sensor (e.g., camera) data, that identify an individual based at least in part on the heat signature in an environment having a different (e.g., lower) heat signature. The sensor used at least in part to identify the bodily location (e.g., forehead) of the individual can be the same or different from the sensor used to measure the bodily characteristic (e.g., temperature) of the individual.
[0220] In some embodiments, the bodily characteristic(s) (e.g., temperature) is measured at a lateral distance from the sensor (e.g., at a focal distance from the sensor). The person can be disposed at about the lateral distance of the measurements. The lateral distance may be at least about 0.1 meter (m), 0.3 m, 0.6 m, 0.9 m, or 1 m. The lateral distance of the measurements from the sensor can be between any of the aforementioned distances (e.g., from about 0.1 m to about 1 m). The sensor may have a dwell time during which the bodily characteristic may be measured. The dwell time may be at most about 0.25 seconds (s), 0.5 s, 1 s, 1.5 s, 2 s, 3 s, 4 s, or 5 s. The dwell time may be of any value between the aforementioned values (e.g., from about 0.25 s to about 5 s). The thermal characteristic (e.g., temperature) can be measured with an accuracy of at least about +/−0.7° C., 0.5° C., 0.25° C., or a higher accuracy. The sensor (e.g., as part of a camera) utilized for the measurement may have a horizontal and vertical fields of view. The vertical field of view may be of at least about 30 degrees (°), 40°, 55°, 60°, 70°, 80°, 90°, 100°, 110°, 120°, or 150°. The horizontal field of view may be of at least about 30°, 40°, 55°, 60°, 65°, 70°, 75°, 80°, 90°, or 100°. The vertical field of view may be larger than the vertical field of view. Larger may be by about 1.1 times (*), 1.2*, 1.4*, 1.5*, 1.7*, 1.9*, or 2.0*.
[0221] At times, a plurality of sensors may be utilized to view the bodily characteristic. The density of measurements over time of the plurality of sensors may be (e.g., substantially) similar. For example, a thermal IR camera and a depth camera may be utilized to ascertain a user's bodily location (e.g., forehead) and/or measure the bodily characteristics (e.g., temperature). The sensor(s) may be disposed at a location that facilitates focusing on the bodily location (e.g., withing the horizontal and/or vertical field of view of the user). The sensor(s) may be operatively coupled to actuator(s) that facilitate its translation (e.g., horizontally and/or vertically) to adjust capturing the bodily location of the user within a field of view(s) of the sensor. The sensor(s) may be stationary. The translation may be manual or automatic (e.g., using at least one controller).
[0222]
[0223] In some embodiments, evaluating bodily characteristic(s) of an individual comprises a plurality of operations. For example, the evaluation can comprise identification of facial landmarks of the user and/or distance of the user from the sensor(s) (e.g., actual and/or expected). A control system may direct the sensor(s) to focus of a bodily location (e.g., based at least in part on the facial landmark recognition and/or distance of the user from the sensor(s)). Once the sensor(s) is configured to focus on the bodily location, the sensor(s) acquire measurement(s) from the bodily location. The measurements may be processed (e.g., adjusted according to various adjustment methodologies) to generate a result of the measured bodily characteristic(s). The result may be compared with a threshold (e.g., value), and a report may be generated. The report may be sent to the user, to authorities, to management, or any combination thereof. The report may contain the adjusted bodily characteristic(s) (e.g., to reflect real value of the bodily characteristic(s)), variance from normal value, and any remedial and/or promotional measure actions (e.g., suggestions and/or directions). The report may be utilized as disclosed herein. The non-manipulated temperature measurements, the processed results, the comparison with the threshold, and/or the report may be saved in one or more databases. The databases may be operatively coupled to the network. The data of the database may be utilized to increase the accuracy of the reported. In some embodiments, the measurements of the bodily characteristic(s) are performed in a contactless manner.
[0224] In some embodiments, the raw data of the bodily characteristics of a user measured by the sensor does not accurately reflect the actual bodily characteristics of the user. The raw data require adjustment. The adjustment may be performed automatically (e.g., by a processor). The manner of adjustment may be suggested by an artificial intelligence (AI) computational scheme. The AI may comprise a learning module. The learning module may utilize historic measurements of individuals and/or objects having the subject characteristic (e.g., temperature), as compared to a ground truth (e.g., a local thermometer providing an accurate reading). The learning module may utilize synthetic measurements as part of its learning set. The learning module may utilize simulation as part of its learning set. For example, an IR sensor measuring temperature of an individual at a distance at time t, can be compared to a thermometer measuring the individual's temperature at the time t. The adjustment may consider modeling (e.g., physics modeling) that emulate the subject characteristic measured by the sensors. For example, when the characteristic is temperature, black body radiation at a position of an average individual forehead may be simulated and fed to the learning module as a synthetic measurement as part of its learning set. The learning module (including machine learning) may utilize regression and/or classification algorithms. An output of the machine learning module may be a (e.g., lateral) distance of the subject from the sensor(s) and/or the bodily characteristic (e.g., temperature) measured.
[0225] In some embodiments, the bodily characteristic is measured using a sensor array (e.g., a camera having a sensor array such as an IR or a visible light camera). The sensor array may comprise at least about 25 pixels, 32 pixels, 800 pixels, 1080 pixels, 1280 pixels, or 1920 pixels in its FLS. (e.g., the array may be 25×25, 25×32, or 32×32 sensor array). The sensor array may comprise at least about 1 megapixels (Mpxl), 4 Mpxl, 8 Mpxl, 10 Mpxl, or 12 Mpxl at its FLS. The sensor may measure a wide operational range, e.g., wide span of the bodily characteristic. For example, the temperature sensor may be configured to measure a temperature spanning from about −40° C. to about 85° C. operational temperature range, or from about −40° C. to about 300° C. operational temperature range. The sensor may be a high accuracy sensor. For example, the temperature sensor may have an accuracy of at least about ±1° C., or ±0.5° C. (or any other temperature accuracy value disclosed herein). The sensor array (e.g., camera having the sensor array) may have an adjustable focus (e.g., automatically adjustable using at least one controller). The camera may have one or more lenses. The sensor array may be sensitive to at least the visible spectrum (e.g., comprise an RGB sensor). The sensor array may be sensitive to at least the infrared spectrum. The sensor may be included in a camera comprising stereo vision. The sensor may be part of a camera. The camera may have a shutter (e.g., a rolling shutter). The sensor may have a small pixel size. The small pixel size may have a FLS of at most about 1.2 micrometers (μm), 1.4 μm, 1.5 μm, 2.0 μm, 2.5 μm, or 3 μm.
[0226] In some embodiments, the sensor may be free to move. Movements of the sensor may be controller (e.g., automatically by at least one controller). Movements of the sensor may be effectuated by an actuator operatively coupled to the sensor. The sensor may be configured to have at least 1, 2, 3, 4, 5, or 6 degrees of freedom. The six degrees of freedom may comprise forward, backward, up, down, left, right, pitch, yaw, or roll.
[0227] In some embodiments, a depth camera is utilized to filter noise, and/or to differentiate the user from a background. The camera may comprise stereo vision. The camera may comprise a plurality of sensors spaced apart, e.g., configured to enable the stereo vision capability. The camera module may comprise a processor. The depth camera may include a sensor (e.g., as part of a sensor array) sensitive to the visible spectrum (e.g., Red Green Blue (RGB) sensor), and/or a rolling shutter. The depth camera module may integrate data from IR and visible sensors. The depth camera may comprise IR sensor, IR laser, or a visible sensor. For example, the depth camera may comprise a plurality of visible sensors and one IR sensor and/or laser. For example, the depth camera may comprise a plurality of IR sensors and/or lasers and one visible sensor. The depth camera may comprise a laser (e.g., IR laser). The depth camera may be a web-camera. The depth camera may project and/or sense infrared radiation. The depth camera may utilize comparison of captured data (e.g., sensed information) from two sensors spaced (e.g., horizontally and/or vertically) apart from each other.
[0228]
[0229]
[0230] The at least one processor may comprise a single circuit board computer (SBC). The at least one processor may be configured to run a plurality of neural networks in parallel (e.g., for image classification, object detection, segmentation, and/or speech processing). The at least one processor may be powered by at most about 10 watts (W), 8 W, 5 W, or 4 W.
[0231] In some embodiments, a framing station is utilized to provide an interactive experience with a user.
[0232] In some embodiments, the framing system includes internal framing portions that are integrated together to a single framing system. The framing system can include cabling and one or more interactive devices. The one or more interactive device may include dispensers, sensors, and/or emitters. Sensors can include optical (e.g., electromagnetic) sensors. The emitters may comprise lighting, projector, or media display. The internal framing portions may be integrated with each other in one or more various joint types. The joint types may comprise linear joints or non-linear (e.g., staggered) joints. The joints may comprise butt, dovetail, mitered, mortise-and-tanon, biscuit, picket, rabbet, scarf joint, V-joint, lap joint, strap joint, or tongue-and-groove joint. The butt joint may comprise a simple butt joint, or double butt lap joint. The lap joint may comprise a double butt lap joint, half lap, plain lap joint, beveled lap joint, double lap joint, or joggle lap joint. The strap joint may comprise a single strap joint, double strap joint, recessed double strap joint, or beveled double strap joint. The tongue and groove joint may comprise a landed scarf tongue and groove joint.
[0233] In some embodiments, the framing system (also herein referred to as “framing apparatus”) comprises a casing configured to house one or more processors and/or wiring. The casing may comprise one or more openings. The opening(s) may facilitate operatively coupling (e.g., wired and/or wirelessly connecting) the framing system to the network (e.g., local network of the facility). The opening(s) may facilitate servicing one or more components of the framing system disposed in the casing. For example, the opening may facilitate servicing the wiring and/or processor(s) disposed in the casing. Servicing comprises maintenance, replacement, or introduction of new components. The openings may comprise a lid or door that is reversibly openable and closeable. The lid or door may be fastened to the body of the casing by one or more hingers or screws. The lid or door may comprise a mechanism that facilitates snapping the lid or door (respectively) to the body of the casing. An opening of the casing may be disposed at the front, back, or side of the casing, with the front being a side designed to be confronted by a user interacting with the interactive framing system. For example, framing systems 2900 depicts the front side of the framing system, and framing system 2940 depicts its back side view. At times, the interactive framing system is designed for interaction on both its opposing side with a plurality of users. In that case the lid or door of the casing may be at one side designed for interacting with a user, or at its opposing side designed for interacting with another user (e.g., whether simultaneously or non-simultaneously). An internal space of the framing portions (that form the framing system) may be utilized to house wiring and/or device(s) (e.g., sensors or emitters. E.g., projector 342 of
[0234]
[0235]
[0236]
[0237]
[0238]
[0239]
[0240] In some embodiments, the interactive framing system is configured to service user(s) on one of its sides. In some embodiments, the interactive framing system is configured to service user(s) on both of its sides.
[0241] While preferred embodiments of the present invention have been shown, and described herein, it will be obvious to those skilled in the art that such embodiments are provided by way of example only. It is not intended that the invention be limited by the specific examples provided within the specification. While the invention has been described with reference to the afore-mentioned specification, the descriptions and illustrations of the embodiments herein are not meant to be construed in a limiting sense. Numerous variations, changes, and substitutions will now occur to those skilled in the art without departing from the invention. Furthermore, it shall be understood that all aspects of the invention are not limited to the specific depictions, configurations, or relative proportions set forth herein which depend upon a variety of conditions and variables. It should be understood that various alternatives to the embodiments of the invention described herein might be employed in practicing the invention. It is therefore contemplated that the invention shall also cover any such alternatives, modifications, variations, or equivalents. It is intended that the following claims define the scope of the invention and that methods and structures within the scope of these claims and their equivalents be covered thereby.