BI-LEVEL OPTIMIZATION-BASED INFRARED AND VISIBLE LIGHT FUSION METHOD
20220044442 · 2022-02-10
Inventors
- Risheng LIU (Dalian, Liaoning, CN)
- Xin FAN (Dalian, Liaoning, CN)
- Jinyuan LIU (Dalian, Liaoning, CN)
- Wei ZHONG (Dalian, Liaoning, CN)
- Zhongxuan LUO (Dalian, Liaoning, CN)
Cpc classification
G06T7/80
PHYSICS
International classification
G06T7/80
PHYSICS
Abstract
The present invention proposes a bi-level optimization-based infrared and visible light fusion method, adopts a pair of infrared camera and visible light camera to acquire images, and relates to the construction of a bi-level paradigm infrared and visible light image fusion algorithm, which is an infrared and visible light fusion algorithm using mathematical modeling. Binocular cameras and NVIDIA TX2 are used to construct a high-performance computing platform and to construct a high-performance solving algorithm to obtain a high-quality infrared and visible light fusion image. The system is easy to construct, and the input data can be acquired by using stereo binocular infrared and visible light cameras respectively; the program is simple and easy to implement; and the fusion image is divided into an image domain and a gradient domain for fusion by means of mathematical modeling according to different imaging principles of infrared and visible light cameras.
Claims
1. A bi-level optimization-based infrared and visible light fusion method, wherein the method comprises the following steps: 1) obtaining registered infrared and visible light images, and respectively calibrating each lens and jointly calibrating the respective systems of the visible light binocular camera and the infrared binocular camera; 1-1) respectively calibrating the infrared camera and the visible light camera by the Zhangzhengyou calibration method to obtain internal parameters including focal length and principal point position and external parameters including rotation and translation of each camera; 1-2) calculating the positional relationship of the same plane in the visible light image and the infrared image by using RT obtained by joint calibration and the detected checker corners, and registering the visible light image to the infrared image by using a homography matrix; 2) converting the color space of the visible light image from an RGB image to an HSV image, extracting the value information of the color image as the input of image fusion, and retaining the original hue and saturation; 3) carrying out bi-level paradigm-based mathematical modeling on the input infrared image and the visible light image with the color space converted; establishing two separate models, namely the upper-level sub-problem Leader and the lower-level sub-problem Follower to solve the same problem.
F.sup.k+1=α.sup.kF.sub.l.sup.k+1+(1−α.sup.k)F.sub.f.sup.k+1 wherein F.sup.k+1 is the final result of each iteration and is manually selected based on experience; 7) converting the color space: converting the fusion image back to the RGB image, and adding the hue and saturation previously retained; restoring to the RGB color space from the HSV color space by updating the V information saved into the fusion image in combination with the previously retained H and S; 8) enhancing the color: enhancing the color of the fusion image to generate a fusion image with higher resolution and contrast; and performing pixel-level image enhancement for the contrast of each pixel; performing color correction and enhancement on the restored image to generate a three-channel image that is consistent with observation and detection; and performing color enhancement on the R channel, G channel and B channel respectively to obtain the final fusion image.
2. The bi-level optimization-based infrared and visible light fusion method according to claim 1, wherein the color space conversion of the visible light image in step 2) comprises: 2-1) converting the RGB color space to the HSV color space, wherein V is value, H is hue and S is saturation; and extracting the value information of the visible light image to be fused with the infrared image, and retaining the hue and saturation, wherein the specific conversion is shown as follows:
R′=R/255 G′=G/255 B′=B/255
Cmax=max(R′,G′,B′)
Cmin=min(R′,G′,B′)
Δ=Cmax−Cmin
V=Cmax 2-2) extracting the V channel as the input of visible light, retaining H and S to the corresponding matrix, and retaining the color information for the subsequent color restoration after fusion.
3. The bi-level optimization-based infrared and visible light fusion method according to claim 1, wherein the solving method in step 5): first introducing two auxiliary variables u, w, wherein u=∇F−∇I, w=∇F−∇V; and through variable substitution, transforming to minimizing the following problem:
λ.sub.1.sup.k+1=λ.sub.1.sup.k−ρ.sub.1(∇F−∇V−w)
λ.sub.2.sup.k+1=λ.sub.2.sup.k−ρ.sub.2(∇F−∇I−u).
4. The bi-level optimization-based infrared and visible light fusion method according to claim 1, wherein the specific formulas for color space conversion in step 7) are shown as follows:
R′,G′,B′=((R′+m)×255,(G′+m)×255,(B′+m)×255) wherein C is the product of the value and the saturation; and in is the difference of the value and C.
5. The bi-level optimization-based infrared and visible light fusion method according to claim 1, wherein the specific formulas for color enhancement in step 8) are shown as follows:
R.sub.out=(R.sub.in).sup.1/gamma
R.sub.display=(R.sub.in.sup.(1/gamma)).sup.gamma
G.sub.out=(G.sub.in).sup.1/gamma
G=(G.sub.in.sup.(1/gamma))gamma
B.sub.out=(B.sub.in).sup.1/gamma
B.sub.display=(B.sub.in.sup.(1/gamma)).sup.gamma wherein gamma is the correction parameter, R.sub.in, G.sub.in and B.sub.in are the values of the three input channels R, G, and B respectively, R.sub.out, G.sub.out and B.sub.out are the intermediate parameters, and R.sub.display, G.sub.display and B.sub.display are the values of the three channels after enhancement.
Description
DESCRIPTION OF DRAWINGS
[0037]
[0038]
DETAILED DESCRIPTION
[0039] The present invention proposes a method for real-time image fusion by an infrared camera and α visible light camera, and will be described in detail below in combination with drawings and embodiments.
[0040] The binocular stereo cameras are placed on a fixed platform, the image resolution of the experiment cameras is 1280×720, and the field of view is 45.4°. To ensure real-time performance, NVIDIA TX2 is used for calculation. On this basis, a real-time infrared and visible light fusion method is designed, and the method comprises the following steps:
[0041] 1) Obtaining registered infrared and visible light images;
[0042] 1-1) Respectively calibrating each lens and jointly calibrating the respective systems of the visible light binocular camera and the infrared binocular camera;
[0043] 1-2) Respectively calibrating the infrared camera and the visible light camera by the Zhangzhengyou calibration method to obtain internal parameters such as focal length and principal point position and external parameters such as rotation and translation of each camera.
[0044] 1-3) Calculating the positional relationship of the same plane in the visible light image and the infrared image by using RT obtained by joint calibration and the detected checker corners, and registering the visible light image to the infrared image by using a homography matrix.
[0045] 2) Converting the color space of the image
[0046] 2-1) In view of the problem that the visible light image has RGB three channels, converting the RGB color space to the HSV color space, extracting the V (value) information of the visible light image to be fused with the infrared image, and retaining H (hue) and S (saturation), wherein the specific conversion is shown as follows:
R′=R/255 G′=G/255 B′=B/255
Cmax=max(R′,G′,B′)
Cmin=min(R′,G′,B′)
Δ=Cmax−Cmin
V=Cmax
[0047] 2-2) Extracting the V (value) channel as the input of visible light, retaining H (hue) and S (saturation) to the corresponding matrix, and retaining the color information for the subsequent color restoration after fusion.
[0048] 3) Carrying out bi-level paradigm-based mathematical modeling for the input infrared image and the visible light image with the color space converted. The core of the idea is to establish two separate models, namely the upper-level sub-problem Leader and the lower-level sub-problem Follower to solve the same problem. First, the objective function and the problem constraint in optimization are assumed to be two participants in the game, wherein the objective function is regarded as the Leader and the problem constraint is the Follower. In this kind of competition, the next optimization of the Leader needs to consider the result of the Follower. The two competitors are defined as two composite minimization sub-problems:
[0049] wherein F represents the fused image, and the infrared image and the visible light image are respectively represented by I, V, ∇ represents the operator for obtaining the gradient, and γ, β represent parameters of Leader and Follower respectively.
[0050] 4) Solving the upper-level sub-problem; the target result can be obtained by solving the following formula:
[0051] wherein F.sub.1.sup.k−1 represents the result of the upper-level problem. The goal has a simple closed-form solution and thus can be directly obtained by the closed-form solution shown as the following formula:
[0052] 5) Solving the lower-level sub-problem; it can be found that the problem is non-convex and non-smooth and difficult to solve directly and is solved by the alternating direction multiplier method widely used. Therefore, it is necessary to transform an unconstrained problem into a constrained problem by introducing auxiliary variables, and then the problem is solved under the framework. Specifically, first introducing two auxiliary variables u, w, wherein u=∇F−∇I, w=∇F−∇V. Through variable substitution, transforming to minimizing the following problem:
[0053] wherein ∇ represents the gradient operator, λ.sub.1, λ.sub.2 are two multipliers, ρ.sub.1, ρ.sub.2 are parameters of a penalty term, and three sub-problems respectively about u, w, F are obtained through variable separation:
[0054] 5-1) For the update of F.sub.f.sup.k+1 of the lower-level problem, the closed-form solution of the formula is used to obtain:
[0055] 5-2) The multipliers λ.sub.1, λ.sub.2 need updating after each iteration, and the specific update mode is as follows:
λ.sub.1.sup.k+1=λ.sub.1.sup.k−ρ.sub.1(∇F−∇V−w)
λ.sub.2.sup.k+1=λ.sub.2.sup.k−ρ.sub.2(∇F−∇I−u)
[0056] 6) Obtaining two estimates F.sub.1.sup.k+1, F.sub.f.sup.k+1 of the fusion result under different characteristics by solving the upper-level and lower-lever sub-problems, and to fuse the two components into an image F, linearly combining the two components, which is expressed as the following form:
F.sup.k+1=α.sup.kF.sub.l.sup.k+1+(1−α.sup.k)F.sub.f.sup.k+1
[0057] wherein F is the final result of each iteration, and α is a parameter weighing the two components. The parameter needs to be manually selected based on experience, and is selected as 0.5 herein.
[0058] 7-1) Restoring to the RGB color space from the HSV color space by updating the V (value) information saved into the fusion image in combination with the previously retained H (hue) and S (saturation), wherein the specific formulas are shown as follows:
R′,G′,B′=((R′+m)×255,(G′+m)×255,(B′+m)×255)
[0059] wherein C is the product of the value and the saturation; and in is the difference of the value and C.
[0060] 7-2) performing color correction and enhancement on the image restored in step 7-1 to generate a three-channel image that is consistent with observation and detection; and performing color enhancement on the R channel, G channel and B channel respectively, wherein the specific formulas are shown as follows:
R.sub.out=(R.sub.in).sup.1/gamma
R.sub.display=(R.sub.in.sup.(1/gamma)).sup.gamma
G.sub.out=(G.sub.in).sup.1/gamma
G=(G.sub.in.sup.(1/gamma)).sup.gamma
B.sub.out=(B.sub.in).sup.1/gamma
B.sub.display=(B.sub.in.sup.(1/gamma)).sup.gamma
[0061] wherein gamma is the correction parameter, R.sub.in, G.sub.in and B.sub.in are the values of the three input channels R, G, and B respectively, R.sub.out, G.sub.out and B.sub.out are the intermediate parameters, and R.sub.display, G.sub.display and B.sub.display are the values of the three channels after enhancement.