METHOD AND SYSTEM FOR OPTICAL MONITORING OF UNMANNED AERIAL VEHICLES BASED ON THREE-DIMENSIONAL LIGHT FIELD TECHNOLOGY
20220210375 · 2022-06-30
Inventors
Cpc classification
G06T3/4053
PHYSICS
H04N23/00
ELECTRICITY
H04N13/254
ELECTRICITY
G06V30/2504
PHYSICS
G06V20/41
PHYSICS
G06V20/647
PHYSICS
H04N13/239
ELECTRICITY
H04N13/282
ELECTRICITY
H04N7/18
ELECTRICITY
H04N2013/0081
ELECTRICITY
International classification
H04N7/18
ELECTRICITY
G06T3/40
PHYSICS
Abstract
Disclosed in the present invention are a method and a system for monitoring unmanned aerial vehicles based on three-dimensional light field technology. Provided is an unmanned aerial vehicle monitoring method based on three-dimensional light field technology, comprising: beginning unmanned aerial vehicle monitoring; by means of a light field camera, acquiring low resolution video image information; determining whether the acquired video image information is an unmanned aerial vehicle; performing graphic reconstruction on an unmanned aerial vehicle image therein; and acquiring reconstructed light field image depth and position information to monitor the unmanned aerial vehicle and emitting an alert. The method and the system for monitoring in the present invention are able to acquire a clear stereoscopic image, thus raising efficiency and accuracy in the process of unmanned aerial vehicle monitoring or detection.
Claims
1. A method of monitoring an unmanned aerial vehicle based on light field technology, comprising: starting the procedure of unmanned aerial vehicle monitor; obtaining video image information having low resolution by means of a light field camera; determining whether the acquired video image information shows the unmanned aerial vehicle; reconstructing graphics of an unmanned aerial vehicle image therein; and acquiring depth and position information of the reconstructed light field image to monitor the unmanned aerial vehicle and sending an alert.
2. The method of monitoring an unmanned aerial vehicle of claim 1, wherein the graphics reconstruction step is a super-resolution method based on a model or a super-resolution method based on multiple frames.
3. The method of monitoring an unmanned aerial vehicle of claim 2, wherein the super-resolution method based on multiple frames is a method for reconstructing a high-resolution (HR) image from a group of relatively low-resolution (LR) images, which can also be referred to as a multi-frame image super-resolution method; the reconstruction step is performed by using the correlation between the recorded relatively low-resolution (LR) images.
4. The method of monitoring an unmanned aerial vehicle of claim 1, wherein the step of reconstructing the unmanned aerial vehicle image comprises (a) a light field image generated by a micro-lens array; (b) a sequence of sub-aperture images arranged according to a distance of a focal plane; (c) a single sub-aperture image; and (d) multi-view-angle sub-aperture images arranged according to positions on a main lens; wherein the multi-view-angle sub-aperture image array is obtained after an original compound eye image is processed.
5. The method of monitoring an unmanned aerial vehicle of claim 1, wherein after the step of reconstructing the unmanned aerial vehicle image, the reconstructed light field image is combined with an estimated depth map to refocus the light field images, and the refocused images are combined to obtain all focused light field images.
6. The method of monitoring an unmanned aerial vehicle of claim 4, wherein according to synthetic aperture technology, pixel points in the original compound eye image are re-projected into each sub-aperture image, such that a scenario is imaged from different view angles, light field information in the original compound eye image is further synthesized and extracted, a multi-view-angle view of an imaging space is obtained, and a digital refocusing sequence and a depth map are obtained.
7. The method of monitoring an unmanned aerial vehicle of claim 6, wherein the following formulas are adopted:
8. The method of monitoring an unmanned aerial vehicle of claim 1, further comprising: classifying image frames of the video image; segmenting semantics on the image frames; detecting the unmanned aerial vehicle image; and segmenting instances on the unmanned aerial vehicle image to cluster pixels of different objects.
9. The method of monitoring an unmanned aerial vehicle of claim 1, further comprising: analyzing a super-resolution frame of the video image after processing and decoding the light field image; remolding a frame sequence in the video into a data matrix, for a given preprocessed monitoring video; analyzing and encoding the data matrix to extract main features of an image; and identifying an unmanned aerial vehicle image feature pattern by utilizing machine learning technology, and then detecting the unmanned aerial vehicle in the video image.
10. The method of monitoring an unmanned aerial vehicle of claim 9, wherein the machine learning technology is RCNN, Fast RCNN, YOLO or SSD method.
11. A system of monitoring an unmanned aerial vehicle based on light field technology, comprising: a starting module configured to begin monitoring an unmanned aerial vehicle; an image information acquisition module configured to, by means of a light field camera, acquire video image information having low resolution; a judgment module configured to determine whether the acquired video image information shows the unmanned aerial vehicle; a reconstruction module configured to perform graphic reconstruction on an unmanned aerial vehicle image therein; and an alerting module configured to acquire depth and position information of the reconstructed light field image to monitor the unmanned aerial vehicle and send an alert.
12. The system of monitoring an unmanned aerial vehicle of claim 11, wherein the image reconstruction step is a super-resolution method based on a model or a super-resolution method based on multiple frames.
13. The system of monitoring an unmanned aerial vehicle of claim 12, wherein the super-resolution method based on multiple frames is a method for reconstructing a high-resolution (HR) image from a group of relatively low-resolution (LR) images, which can also be referred to as a multi-frame image super-resolution method; the reconstruction step is performed by using correlation between the recorded relatively low-resolution (LR) images.
14. The system of monitoring an unmanned aerial vehicle of claim 11, wherein the step of reconstructing the unmanned aerial vehicle image comprises (a) a light field image generated by a micro-lens array; (b) a sequence of sub-aperture images arranged according to a distance of a focal plane; (c) a single sub-aperture image; and (d) multi-view-angle sub-aperture images arranged according to positions on a main lens; wherein the multi-view-angle sub-aperture image array is obtained after an original compound eye image is processed.
15. The system of monitoring an unmanned aerial vehicle of claim 11, wherein after the step of reconstructing the unmanned aerial vehicle image, the reconstructed light field image is combined with an estimated depth map to refocus the light field images, and the refocused images are combined to obtain all focused light field images.
16. The system of monitoring an unmanned aerial vehicle of claim 13, wherein according to synthetic aperture technology, pixel points in the original compound eye image are re-projected into each sub-aperture image, such that a scenario is imaged from different view angles, light field information in the original compound eye image is further synthesized and extracted, a multi-view-angle view of an imaging space is obtained, and a digital refocusing sequence and a depth map are obtained.
17. The system of monitoring an unmanned aerial vehicle of claim 15, wherein the following formulas are adopted:
18. The system of monitoring an unmanned aerial vehicle of claim 11, further comprising: a classification module configured to classify image frames of the video image; a segmentation module configured to perform semantic segmentation on the image frames; and a detecting module configured to detect the unmanned aerial vehicle image, and wherein instance segmentation is performed on the unmanned aerial vehicle image to cluster pixels of different objects.
19. The system of monitoring an unmanned aerial vehicle of claim 11, further comprising: an analysis module configured to analyze a super-resolution frame of the video image after the light field image is processed and decoded; a remolding module configured to, for a given preprocessed monitoring video, remold a frame sequence in the video into a data matrix; a feature extraction module configured to analyze and encode the data matrix to extract main features of an image; and an identification module configured to identify an unmanned aerial vehicle image feature pattern by utilizing machine learning technology, and then detect the unmanned aerial vehicle in the video image.
20. The system of monitoring an unmanned aerial vehicle of claim 19, wherein the machine learning technology is RCNN, Fast RCNN, YOLO or SSD method.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0009] To describe the technical solutions in the embodiments of the present invention clearly, the following contents briefly describe the accompanying drawings required for the embodiments. Apparently, the accompanying drawings in the following description show merely some examples of the present invention, and a person of ordinary skill in the art may still derive other drawings from these accompanying drawings without innovative efforts.
[0010]
[0011]
[0012]
[0013]
[0014]
[0015]
[0016]
[0017]
[0018]
[0019]
[0020]
[0021]
DETAILED DESCRIPTION
[0022] Embodiments of the present invention will now be described with reference to the accompanying drawings. The present invention may, however, be implemented in many different forms and should not be construed as being limited to the embodiments set forth herein. These embodiments are provided only for thoroughness and completeness of the present invention, such that the scope of the present invention may be fully conveyed to those skilled in the art. The terms used in the detailed description of the embodiments illustrated in the accompanying drawings are not intended to limit the present invention.
[0023]
[0024]
[0025]
wherein I and I′ represent coordinate systems of a primary imaging surface and a secondary imaging surface; and
L and L′ represent energy of the primary imaging surface and the secondary imaging surface.
[0026] After acquisition of depth data of a photographed object in each micro-lens, the depth map of the whole image may be calculated, and thus a 3D video image may be shot.
[0027] At the above-mentioned step 104, the high-resolution image and the light field information of the target unmanned aerial vehicle is obtained by using the light field camera. Different from conventional cameras, the light field camera captures not only two-dimensional images, but also the direction of incident light rays.
[0028] A light field image I(x,y) may be represented by the formula:
I(x,y)=∫∫L.sub.F(u,v,x,y)dudv (1)
wherein (u,v,x,y) represents light traveling along a light ray intersecting the main lens at (u,v) and a micro-lens plane at (x,y), and full aperture is used.
[0029] A shifted light field function may be represented as:
[0030] Light field imaging technology allows refocusing the image and estimating a depth map of a scenario. A basic depth range is computed by means of a light field, and the position of the unmanned aerial vehicle is determined by combining the depth range with a position on the image.
[0031] For fabrication of semiconductors applied on a chip board, a compound eye may be used to find a maximum loop height of an aluminum bonding wire, a first bonding height on a chip and a second bonding height on a substrate.
[0032] In the present invention, the spatial position of the monitored unmanned aerial vehicle is located by using the light field imaging technology. First, no proper focusing operation is required before photography. A post-focusing operation may be performed after the image is captured. Second, all depth map information may be captured by using only one shoot. Three-dimensional information of the position of any object appearing on the shot image may be determined by using the depth map information.
[0033] A distance between the unmanned aerial vehicle and a camera may be measured by using a distance/depth calculation function. Light field technology and structured light vision technology are mainly utilized in the measurement process. A light field is a function which may describe the amount of optical radiation transmitted through each aspect of each point in a space (that is, describe radiative transmission characteristics of light in a three-dimensional space). All useful information in the scenario may be captured by utilizing light field photography technology. The light field camera (also known as a full-aperture camera) may capture both information relevant to an intensity of light in the scenario and information relevant to a direction of propagation of the light in the space. A micro-lens array serves as a core component of the light field camera. Each micro-lens covers a plurality of optical sensor pixels and may separate light radiated to the micro-lens to form a small image on the underlying pixel. The application of the micro-lens array in imaging is technology inspired by animal vision and therefore, the micro-lens array is also called “compound eye” in the light field camera. By combining the main lens with the compound eye, the 4D light field information may be recorded on the optical sensor, and the light field image decoding process is further realized. A functional module mainly includes the following three aspects: image reconstruction, refocusing and depth map estimation. As a basis of light field decoding, the image reconstruction mainly involves ray tracking of a 4D light field; the refocusing is essentially one example of the synthetic aperture technology; the depth map obtained by estimation from the above two steps is always a portal into a three-dimensional world. After the light field image is decoded, the distance between the target and the camera may be calculated.
[0034]
[0035]
[0036]
[0037] After the unmanned aerial vehicle is monitored and identified, the camera is aligned with the unmanned aerial vehicle and refocused to capture a close shot photograph thereof so as to acquire more relevant information. Since the unmanned aerial vehicle always moves in the air, the camera should be tilted towards the direction of the target to form high-quality images. Therefore, image tracking technology is utilized in the system to lock a region of the unmanned aerial vehicle, and intelligent analysis is performed to predict a flight path of the unmanned aerial vehicle. After the unmanned aerial vehicle is detected and tracked, a target scenario and the flight path are transmitted to a motion control unit. The camera is driven by a hexapod robot to be stably aligned with the unmanned aerial vehicle.
[0038] Software of the system manages all the functional modules to work together. The monitoring video is automatically analyzed by the system to detect and track the unmanned aerial vehicle. The light field technology is used for calculation of the target distance and refocusing for the photograph scenario. Resolution reconstruction technology guarantees the imaging quality of all processes. The algorithm involves digital image enhancement, light field decoding, pattern recognition, and machine intelligence. Software and hardware which are coordinated and interact with each other implement a high-precision and full-automatic unmanned aerial vehicle monitoring system.
[0039]
[0040] The “an embodiment”, “one embodiment” or “one or more embodiments” mentioned herein means that the specific features, structures or characteristics described in combination with the embodiment(s) would be included in at least one embodiment of the present invention. Moreover, it should be noted that, the wording “in an embodiment” herein may not necessarily refer to the same embodiment.
[0041] The above description is only for the purpose of illustrating the technical solutions of the present invention, and any person skilled in the art may make modifications and changes to the above embodiments without departing from the spirit and scope of the present invention. Therefore, the protection scope of the present invention should be subject to the claims. The present invention has been described above with reference to examples. However, other embodiments than the above embodiments are equally possible within the scope of this disclosure. The different features and steps of the present invention may be combined in other ways than those described. The scope of the present invention is limited only by the appended claims. More generally, those skilled in the art will readily appreciate that all parameters, dimensions, materials, and configurations described herein are meant to be exemplary and that actual parameters, dimensions, materials, and/or configurations will depend upon specific applications or applications for which the teachings of the present invention are used.