Electronic Monitoring System Using Video Notification
20220345669 · 2022-10-27
Inventors
Cpc classification
H04N7/188
ELECTRICITY
G08B13/19684
PHYSICS
G06V20/52
PHYSICS
G06V40/23
PHYSICS
G08B13/19691
PHYSICS
International classification
Abstract
A camera-based monitoring system is provided that that, upon generation of an alert or notification, can provide a video clip formed from multiple frames or images to the notification system of a user-accessible monitoring device, such as a cell phone, to make it easy detect an object that is moving in the camera's field of view. Since the human eye is extremely sensitive to motion, the “triggering object” whose activities triggered image acquisition can be identified more easily, rapidly, and reliably from the video clip than from a still image. In addition to including the camera and detector(s), the system may include a base station and a controller. A method of operating such an electronic monitoring system also is disclosed
Claims
1. An electronic monitoring system comprising: an imaging device configured to respond to a detected triggering event occurring in an area monitored by the imaging device by capturing visual image frames of the monitored area; and a controller in communication with the imaging device, the controller being configured to execute a program stored in a non-transitory medium and to send a notification to a user-accessible monitoring device upon detection of a triggering event generated by an object, wherein the notification includes a video clip of the monitored area that is formed by or from some of the captured frames.
2. The electronic monitoring system of claim 1, wherein the video clip has a duration of at least three frames
3. The electronic monitoring system of claim 1, wherein the video clip has a duration of from three to 10 frames.
4. The electronic monitoring system of claim 1, wherein the triggering event is motion of a triggering object, wherein the imaging device comprises a video camera having internal memory, and further comprises a motion sensor that is configured to detect motion of the triggering object, wherein the camera is configured, in response to operation of the motion sensor, to capture and transmit a multi-frame video data stream.
5. The electronic monitoring system of claim 4, wherein the system additionally includes a base station communicating with the imaging device and with the monitoring device, and wherein the controller is contained at least in part within the base station.
6. The electronic monitoring system of claim 5, wherein the controller is contained in part in at least one of the imaging device, a sever in communication with the base station and/or the imaging device, and the monitoring device.
7. The electronic monitoring system of claim 1, wherein the controller is configured to determine which of the captured frames is to be contained in or form the basis of the video clip contained in the notification.
8. The electronic monitoring system of claim 7, wherein the controller is configured to select a designated number of sequentially-captured frames captured immediately after detection of the triggering event.
9. The electronic monitoring system of claim 7, wherein the controller is configured to select captured frames determined by the controller to be optimal examples of the detected motion, wherein the captured frames may or may not be sequentially-captured frames.
10. The electronic monitoring system of claim 9, wherein the controller is configured to select captured frames based on detected image characteristics so as to provide an optimal view of the object that triggered the triggering event.
11. An electronic monitoring system comprising: an imaging device including internal memory, a video camera configured to capture and transmit a video image stream, and a motion detector configured to detect movement of a triggering object in an area monitored by the imaging device; a base station in communication with the imaging device; and a controller contained at least in part within at least one of the imaging device and the base station, the controller being configured to, upon detection movement of the triggering object by the motion detector, execute a program stored in a non-transitory medium to send a video clip of the monitored area to a user-accessible monitoring device having a screen capable of displaying video images.
12. The electronic monitoring system of claim 11, wherein the video clip lasts from three to ten frames.
13. A method comprising: detecting a triggering event in an area monitored by an imaging device; in response to detection of the triggering event, using a camera of the imaging device, capturing a series of visual image frames; and displaying a notification on a display of a user-operated monitoring device, the notification comprising a video clip formed by or produced from frames captioned by the camera.
14. The method of claim 13, wherein the video clip has a duration of three to ten frames.
15. The method of claim 13, further comprising selecting, as a basis of the video clip, a designated number of sequentially-captured frames captured after detection of the triggering event.
16. The method of claim 13, further comprising selecting, as a basis of the video clip, a plurality of captured frames that may or may not be sequentially-captured frames.
17. The method of claim 13, wherein the triggering event is movement of an object.
18. The method of claim 17, wherein the selected frames are determined to be optimal examples of detected motion of the moving object.
19. The method of claim 18, wherein the selected frames are selected based on determined image characteristics of the moving object.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0018] Exemplary embodiments of the invention are illustrated in the accompanying drawings in which like reference numerals represent like parts throughout, and in which:
[0019]
[0020]
[0021]
DETAILED DESCRIPTION
[0022] Referring now to
[0023] Each imaging device 12 is configured to acquire data and to transmit it to the base station 14 for further processing and/or transmission to a server and/or the user device(s)16. The acquired data typically will correspond to a video image, and each imaging device 12 may be or include a camera such as a video camera. Several such imaging devices 12 may be mounted around a building or other structure or area being monitored. For example, in the case of a residential home, imaging devices 12 could be mounted by each entrance and selected windows, and even on a gate or light pole. An imaging device 12 also could be incorporated into or coupled to a doorbell, floodlight, etc.
[0024] Still referring to
[0025] Still referring to
[0026] The base station 14 may also be in communication with a sever 58, which may be a cloud-server accessible via the WAN 50. The server 58 can include or be coupled to a microprocessor, a microcontroller or other programmable logic element (individually and collectively considered “a controller”) configured to execute a program. The controller may be contained in whole in the base station 14 or the server 58. Alternatively, interconnected aspects of the controller and the programs executed by it could be distributed in various permutations within the imaging device 12, the base station 14, the monitoring device 16, and the server 58. This program may be utilized in filtering, processing, categorizing, storing, recalling and transmitting data received from the imaging device 12 via the base station 14. Server 58 may also be in communication with or include a computer vision program (“CV”), which can apply one or more filters or processes, such as edge detection, facial recognition, motion detection, etc., to detected one or more characteristics of the recording such as, but not limited to, identifying an individual, animal, vehicle, or package present in the recording. The base station also may be eliminated in its entirety, in which case the imaging device 12 would communicate with the external server 58 via the gateway router 54, and the circuitry in the base station would be contained in whole or in part in the imaging device 12, the gateway router 54, and/or the server 58.
[0027] In operation, each imaging device 12 can be configured, though suitable mounting of the imaging device 12 and/or through suitable manipulation of its controls, to monitor an area of interest, such as a part of a building or section of property. An imaging device 12 may capture an image upon receipt of a command from a monitoring device 16. An image also may be captured automatically upon detection of a triggering event detected by a detector, which may be a standalone detector in communication with the imaging device 12 and/or the base station 14, or which may be on or in the imaging device 12 as illustrated at 21. The triggering event may be motion of a triggering object, and the detector may be a motion detector. Instead of or in addition to detecting motion, the detector could include an IR sensor detecting heat, such as the body heat of an animal or person. The triggering event also could be sound, in which case the detector may include the microphone 18. In this case, the triggering event may be a sound exceeding a designated decibel level or some other identifiable threshold. It is also possible to limit triggering events to detection of a certain class of object, such as an automobile, a person, a package, or an animal Detection of other types of objects would not trigger image capture and recording. Upon receiving notification from an imaging device 12 of a triggering event, the system 10 can generate a push notification (“PN”) or other alert or notification and send it to one or more user devices 16 for indicating the triggering event.
[0028] Whether camera operation is triggered by a command from a monitoring device 16 or by detection of a triggering event by a detector, the camera 24 can then capture a raw video stream which, in turn, can be provided to the media encoder 30 for producing video packets in an encoded video stream. The stream is at least conceptually formed of a number of sequentially-captured frames of the monitored area. Similarly, the microphone 18 and the audio circuit 28 can capture a raw audio stream which, in turn, can be provided to the media encoder 30 for producing audio packets in an encoded audio stream. Accordingly, the video and/or audio packets, referred to herein as “media”packets, are provided in an encoded media stream. Under control of the controller executing the program, the encoded media stream can be transmitted from the wireless I/O communication device 36 to the base station 14.
[0029] The media stream may then be transmitted via the WAN 50 to a remote data storage device in communication with a media server 58 for data storage and processing. The storage device may be a cloud-based storage device, and the media sever 58 may be a cloud server accessible via a wireless connection. A CV may apply one or more filters or processes, such as edge detection, facial recognition, motion detection, etc., to detect one or more characteristics of the recording such as, but not limited to, identifying an individual, animal, vehicle, or package present in the recording. A filtered or otherwise processed image can then be displayed on the monitoring device 16, along with additional visual and/or audio messaging such as a text and/or audio message identifying a generic or particular person or object. The image can be transmitted in the form of frames corresponding to or derived from (via filtering and other processing) of the frames transmitted by the imaging device 12 in the media packets. Alternately, as mentioned above, data storage and processing functions could be performed in combinations of the imaging device 12, the base station 14, and even the monitoring device 16.
[0030] A typical image captured by the imaging device is shown schematically in
[0031] The present invention recognizes that fact that human eyesight is extremely sensitive to motion and can quickly identify a change of a scene over even brief periods of time on the order of a few frames. If the notification displayed on the screen of the monitoring device 16 includes more than one frame in the form of a short video clip or an animated GIF style playback, the user can quickly and easily see what was moving frame by frame and to quickly identify which candidate object is the triggering object that triggered the motion sensor.
[0032] The notification process is shown schematically in the flowchart of
[0033] Next, at Block 86, the controller generates a multi-frame display or “video clip” to be included in the notification that ultimately will be displayed on the screen of the monitoring device 16. The displayed frames could be a number of still images or snapshots or, more typically, a number of frames of a video image. The term “video clip” is used herein as a generic term encompassing both types of displays. The frames forming the video clip could be delivered raw from the imaging device 12. More typically, however, the captured frames forming the basis of the video clip will be filtered and otherwise processed as described above before the video clip is generated and displayed.
[0034] The number of frames in the video clip, and thus the temporal length of the video clip, may vary based on a variety of factors including designer preference and the methodology employed for frame selection. For example, the minimum number of frames to be contained in clip may be driven by the length of the detected triggering event. So, if the triggering event is of a relatively short duration, it would be desirable to select representative frames that omit pre-roll or post object motion frames. As discussed in more detail below, if CV or a more advanced algorithm is used, it may be possible to select only the “best” frames suitable for display. As a practical matter, the clip likely will include at least three frames, with human-eye based motion detection being facilitated by looping the clip as an animated GIF. The maximum number of frames to be included in a clip is limited only by the bandwidth usage the designer wishes to tolerate and the latency for the notification that the designer is willing to accept. Sending more frames than necessary would result in delay of notification receipt by the user until all frames are received by the monitoring device 16. Finally, the operating system employed by the monitoring device could impose bandwidth or other limitations on the number of frames optimally contained in a clip. As a practical matter, a clip length of ten frames or less likely would satisfactorily avoid all of these concerns. A standard GIF is six frames, so a user might consider a six-frame video clip to be particularly familiar
[0035] The particular captured frames to be included in or to form the basis of the video clip may be simply a number of frames that are sequentially captured after the triggering event, typically beginning with the first frame captured after the triggering event. More sophisticated techniques for frame selection are descried below.
[0036] Next, a notification containing the video clip is displayed on the screen of the monitoring device 16 in Block 88. The notification may take the form of a push notification. The notification may show the video clip only once or cycle the clip for so long as the notification is active, such as occurs with the display of a standard animated GIF. The displayed notification could include any of a number of components in addition to the video clip. For example, the notifications could include graphical elements or a text message providing a generic message such as “MOTION DETECTED AT FRONT DOOR” or a more specific message such as “PERSON DETECTED AT REAR DOOR.” The message could be more specific still, such as in identifying a specific type of person “DELIVERY PERSON” or even a specific person or object “BILL” or “BILL'S CAR.” The displayed notification also could include audio aspects, such as a ring or a chime, or even a voice notification conveying the same or different information delivered by text.
[0037] If earliest possible notification is of interest, the process may display a still image on the screen of the user device 16 as soon as motion is detected and an image is captured, and follow with a video image in a brief period of time, typically less than a few seconds, when the GIF or other video is available. Hence, the video clip need not be the only or even the first visual notification displayed by the monitoring device 16 as part of the notification.
[0038] Next, in Block 90, the controller determines whether not the user reacts to the notification by interfacing with the monitoring device 16 within a designated time limit of, for example, 5-30 seconds. If “NO”, the process proceeds to Block 92, where the imaging device 12 stops recording and is placed back into sleep or monitoring mode pending detection of another triggering event. If “YES”, the system 10 may perform any of a number of possible actions selectable by the user, depending on the configuration and capabilities of the particular system. For example, the user may simply instruct the system 10 to ignore the notification, as could be the case, for example, if the triggering object were a swaying tree or a waving flag. In this case, the notification would stop being displayed on user device 16, and the imaging device 12 would stop recording and return to sleep or monitoring mode. The user also could interface with the user device 16 to instruct the imaging device 12 to display an audio or visual alarm, or to simply open a voice communication channel with the imaging device 12. As still another example, the user could initiate a call to an individual or emergency services. In any event, the system 10 will continue to monitoring for another triggering event. Depending on the option selected by the user and the configuration of a particular system, the notification may be turned off, and/or recording may cease.
[0039] As mentioned above, techniques may be employed to optimize the frames displayed in the video clip contained in the notification for triggering object recognition. The captured frames that are selected for the frames included in the video clip may or may not be sequentially captured frames. Any of a number of computer vision and/or AI techniques could be used to select the optimal frames to include in the video clip.
[0040] In one example, the monitored motion can be recorded for a period of time, for example, 10-20 frames, and some of those frames may be selected by the controller to show optimal examples of the detected motion. Computer vision may be employed for this task. For example, the processing software can identify the clearest 3-10 of the first 10-20 frames that are captured upon the triggering event and include only those frames in the video clip. Alternatively, frames in which the triggering object has not moved between frames could be omitted.
[0041] As another example, frame selection could be based on detected image characteristics to provide the best view of the triggering object. For example, if the triggering object is a person, computer vision may be employed to identify the frames that best show the person's face, and the video clip could be limited to those frames.
[0042] Although the best mode contemplated by the inventors of carrying out the present invention is disclosed above, practice of the above invention is not limited thereto. It will be manifest that various additions, modifications and rearrangements of the features of the present invention may be made without deviating from the spirit and the scope of the underlying inventive concept.