VISION SYSTEM FOR A MOTOR VEHICLE
20230171510 · 2023-06-01
Inventors
Cpc classification
G06V10/751
PHYSICS
H04N23/745
ELECTRICITY
G06V20/58
PHYSICS
G08G1/166
PHYSICS
G08G1/09623
PHYSICS
G06V10/36
PHYSICS
International classification
H04N23/745
ELECTRICITY
G06V20/58
PHYSICS
G06V10/36
PHYSICS
Abstract
A vision system (10) for a motor vehicle comprises an imaging apparatus (11) adapted to capture images from a surrounding of the motor vehicle, and a data processing unit (14) adapted to perform image processing on images captured by said imaging apparatus (11) in order to detect objects in the surrounding of the motor vehicle. The data processing unit (14) comprises a flicker mitigation software module (33) adapted to generate a flicker mitigated current image (30′) for a current image frame by filter processing involving a captured current image (30.sub.N+1) corresponding to the current image frame and at least one captured earlier image (30.sub.N) corresponding to an earlier image frame.
Claims
1-15. (canceled)
16. A vision system for a motor vehicle, comprising: a memory; and a processor communicatively coupled to the memory and configured to: receive, from a camera, a first image frame and a second image frame; process the first image frame and the second image frame to detect objects within the first image frame and the second image frame; and generate a flicker mitigated current image based on filter processing the first image frame and the second image frame.
17. The vision system of claim 16, wherein the processor is configured to: detect a light source in the first image frame and the second image frame; and time filter a region around the detected light source in the first image frame and the second image frame.
18. The vision system of claim 17, wherein the processor is configured to blend a first image region around the detected light source in the first image frame with a corresponding second image region in the second image frame.
19. The vision system of claim 18, wherein the processor is configured to blend the first image region with the second image region based on first and second weights.
20. The vision system of claim 19, wherein the first and second weights vary within the first and second image regions.
21. The vision system of claim 20, wherein the first and second weights vary monotonically from a center to an edge of the first and second image regions.
22. The vision system of claim 18, wherein the processor is configured to determine which of the first image region and the second image region has at least one of a higher brightness and pre-defined color.
23. The vision system of claim 17, wherein the processor is configured to blend the second image region around the detected light source in the second image frame over the first image region in the first image frame, wherein the detected light source is visible in the second image frame and not visible in the first image frame.
24. The vision system of claim 17, wherein the processor is configured to track the detected light source over a plurality of image frames comprising the first image frame and the second image frame.
25. The vision system of claim 17, wherein the processor is configured to predict the position of the detected light source in a future image frame.
26. The vision system of claim 16, wherein the processor is configured to: calculate a spatially low pass filtered difference image between the first image frame and the second image frame; and compensate the first image frame based on the spatially low pass filtered difference image.
27. The vision system of claim 26, wherein the processor is configured to calculate the spatially low pass filtered difference image based on a color intensity of the first image frame and the second image frame.
28. The vision system of claim 27, wherein the processor is configured to calculate the spatially low pass filtered difference image between a green pixel intensity of the first image frame and a green pixel intensity of the second image frame.
29. The vision system of claim 16, wherein the camera is configured to capture a plurality of image frames comprising the first image frame and the second image frame at a plurality of exposure settings, and wherein the processor is configured to generate flicker mitigated images from the captured images based on the plurality of exposure settings, the flicker mitigated images comprising the flicker mitigated current image.
30. The vision system of claim 16, wherein the processor is configured to resample the second image frame before the filter processing to compensate for movement of the motor vehicle from a first time associated with the second image frame to a second time associated with the first image frame.
31. The vision system of claim 16, comprising the camera.
32. A method by at least one processor, the method comprising: receiving, from a camera, a first image frame and a second image frame; processing the first image frame and the second image frame to detect objects within the first image frame and the second image frame; and generating a flicker mitigated current image based on filter processing the first image frame and the second image frame.
33. The method of claim 32, comprising: detecting a light source in the first image frame and the second image frame; and time filtering a region around the detected light source in the first image frame and the second image frame.
34. The method of claim 33, comprising blending a first image region around the detected light source in the first image frame with a corresponding second image region in the second image frame.
35. The method of claim 34, comprising blending the first image region with the second image region based on first and second weights.
Description
[0032] In the following the invention shall be illustrated on the basis of preferred embodiments with reference to the accompanying drawings, wherein:
[0033]
[0034]
[0035]
[0036]
[0037]
[0038]
[0039]
[0040]
[0041]
[0042]
[0043] The on-board vision system 10 is mounted, or to be mounted, in or to a motor vehicle and comprises an imaging apparatus 11 for capturing images of a region surrounding the motor vehicle, for example a region in front of the motor vehicle. The imaging apparatus 11, or parts thereof, may be mounted for example behind the vehicle windscreen or windshield, in a vehicle headlight, and/or in the radiator grille. Preferably the imaging apparatus 11 comprises one or more optical imaging devices 12, in particular cameras, preferably operating in the visible wavelength range, or in the infrared wavelength range, or in both visible and infrared wavelength range. In some embodiments the imaging apparatus 11 comprises a plurality of imaging devices 12 in particular forming a stereo imaging apparatus 11. In other embodiments only one imaging device 12 forming a mono imaging apparatus 11 can be used. Each imaging device 12 preferably is a fixed-focus camera, where the focal length f of the lens objective is constant and cannot be varied.
[0044] The imaging apparatus 11 is coupled to an on-board data processing unit 14 (or electronic control unit, ECU) adapted to process the image data received from the imaging apparatus 11. The data processing unit 14 is preferably a digital device which is programmed or programmable and preferably comprises a microprocessor, a microcontroller, a digital signal processor (DSP), and/or a microprocessor part in a System-On-Chip (SoC) device, and preferably has access to, or comprises, a digital data memory 25. The data processing unit 14 may comprise a dedicated hardware device, like a Field Programmable Gate Array (FPGA), an Application Specific Integrated Circuit (ASIC), a Graphics Processing Unit (GPU) or an FPGA and/or ASIC and/or GPU part in a System-On-Chip (SoC) device, for performing certain functions, for example controlling the capture of images by the imaging apparatus 11, receiving the electrical signal containing the image information from the imaging apparatus 11, rectifying or warping pairs of left/right images into alignment and/or creating disparity or depth images. The data processing unit 14 may be connected to the imaging apparatus 11 via a separate cable or a vehicle data bus. In another embodiment the ECU and one or more of the imaging devices 12 can be integrated into a single unit, where a one box solution including the ECU and all imaging devices 12 can be preferred. All steps from imaging, image processing to possible activation or control of a safety device 18 are performed automatically and continuously during driving in real time.
[0045] Image and data processing carried out in the data processing unit 14 advantageously comprises identifying and preferably also classifying possible objects (object candidates) in front of the motor vehicle, such as pedestrians, other vehicles, bi-cyclists and/or large animals, tracking over time the position of objects or object candidates identified in the captured images, and activating or controlling at least one safety device 18 depending on an estimation performed with respect to a tracked object, for example on an estimated collision probability.
[0046] The safety device 18 may comprise at least one active safety device and/or at least one passive safety device. In particular, the safety device 18 may comprise one or more of: at least one safety belt tensioner, at least one passenger air-bag, one or more restraint systems such as occupant airbags, a hood lifter, an electronic stability system, at least one dynamic vehicle control system, such as a brake control system and/or a steering control system, a speed control system; a display device to display information relating to a detected object; a warning device adapted to provide a warning to a driver by suitable optical, acoustical and/or haptic warning signals.
[0047] The invention is applicable to autonomous driving, where the ego vehicle is an autonomous vehicle adapted to drive partly or fully autonomously or automatically, and driving actions of the driver are partially and/or completely replaced or executed by the ego vehicle.
[0048] The problem underlying the present invention is illustrated in
[0049] On order to solve the above problem, the data processing unit 14 comprises a flicker mitigation software module 20 adapted to generate a flicker mitigated current image for a current image frame by filter processing involving a captured current image corresponding to the current image frame and at least one captured earlier image corresponding to an earlier image frame. This is explained in the following for two basic embodiments of the invention. The flicker mitigation software module 20 has access to the data memory 25 where the one or more earlier images needed for the flicker mitigation are stored for use in the current time frame processing.
[0050] A first basic embodiment of the invention is explained with reference to
[0051] A simple practical example of two images 30.sub.N, 30.sub.N+1 corresponding to consecutive time frames N and N+1 is shown in
[0052] By comparing
[0053] The light source detector 31 outputs information relating to the bounding boxes 40, 41, like position and size of these, and the image patches (ROIs) limited by the bounding boxes, to an optional light source tracker 32. The light source tracker 32, if present, is adapted to track the detected light sources over several time frames, and to output corresponding bounding box information 40, 41. For example,
[0054] The light source detector 31 and the light source tracker 32 are software modules similar to conventional object detectors and trackers for detecting and tracking objects like for example other vehicles, pedestrians etc., and may be known per se.
[0055] All information on bounding boxes 40.sub.N, 41.sub.N, 40.sub.N+1, 41.sub.N+1 of consecutive image frames N, N+1, . . . , are forwarded to a flicker mitigation software module 33. The flicker mitigation software module 33 takes the region of interest (ROI) of the traffic light from time frame N (image region in bounding box 40.sub.N and 41.sub.N, respectively), and resamples the ROI of time frame N to the size of the traffic light ROI in the time frame N+1 (image region in bounding box 40.sub.N+1 and 41.sub.N+1, respectively).
[0056] In one embodiment, the flicker mitigation software module 33 calculates an average ROI 40′.sub.N+1, 41′.sub.N+1 from the resampled ROI of time frame N and the ROI of time frame N+1, where calculating an average ROI means calculating an average z value (RGB value, greyscale value or intensity value) of each pixel of the ROI. The flicker mitigation software module 33 then creates a flicker mitigated current image 30′.sub.N+1 by taking the captured current image 30.sub.N+1 everywhere outside the ROIs of detected light sources (here, everywhere outside the ROIs 40.sub.N+1, 41.sub.N+1); while filling in the averaged ROIs 40′.sub.N+1, 41′.sub.N+1 into the bounding boxes of the of the detected light sources.
[0057] As a result, the flicker mitigated current image 30′.sub.N+1 shown in
[0058] In another embodiment, the flicker mitigation software module 33 comprises a brightness and/or color detector which is adapted to detect the brightness and/or color (like green/orange/red in the case of traffic lights) of the detected light sources in the ROIs 40.sub.N, 41.sub.N, 40.sub.N+1, 41.sub.N+1, and to decide which of the ROIs 40.sub.N, 41.sub.N, 40.sub.N+1, 41.sub.N+1 is preferable. In the example of
[0059] In a second basic embodiment of the invention, the flicker mitigation software module 33 is adapted to calculate a spatially low pass filtered difference image between a captured current image 30.sub.N+1 and a captured earlier image 30.sub.N; and preferably to compensate the captured current image 30.sub.N+1 on the basis of the calculated spatially low pass filtered difference image.
[0060] The second basic embodiment of the invention is described in the following with reference to
[0061]
[0062] Before coming to the general case, a simple example with a fairly uniform illumination of the scene will be investigated for a better understanding. Here, the flicker mitigation software module 33 is adapted to calculate the mean (average) of the green pixel intensity (in an RGB color sensor) over every image row of captured images 30 like the one shown in
[0063] The flicker mitigation software module 33 is adapted to calculate the differences between the row mean intensity values (row mean differences) for consecutive frames. The corresponding differences between the row mean intensity values of image frames 1 and 2, frames 2 and 3, frames 3 and 4, and frames 4 and 5 of
[0064] Generalizing the above, the following compensation scheme performed in the flicker mitigation software module 33 is suited for removing the flicker/banding in a perfectly even illuminated scene: [0065] Calculate the green pixel intensity averaged over an image row (row mean) for consecutive frames N+1 and N; [0066] Calculate the row mean difference between frame N+1 and frame N; [0067] spatially low pass filter the row mean difference; [0068] compensate frame N+1 with half of the spatially low pass filtered row mean difference.
[0069] In reality there can be much more varying illumination in a scene. Therefore, instead of calculating one compensation value per row (1D compensation), the flicker mitigation software module 33 should preferably be adapted to perform a 2D compensation. In a similar fashion like above, green pixel intensity differences between two frames are calculated by the flicker mitigation software module 33 in a 2D fashion (instead of 1D). This can be done in several ways, e.g.: [0070] A. Calculate a complete 2D difference image from image N and N+1. Spatially low pass filter it. An example of a complete low pass filtered 2D difference image for the scene of
[0072] When the vehicle is moving, subsequent images N and N+1 capture a slightly different view of the environment since the camera has moved relative to the environment. This can preferably be compensated by resampling image N before calculating the difference image. This will be more computationally efficient when using approach B above compared to approach A, since a lower resolution image, the sub-region image, needs to be resampled compared to resampling the full resolution image.
[0073] The pixel resampling locations can be calculated from, e.g., optical flow or from a model of the environment, or from a combination thereof. The model would use camera calibration and the vehicle movement. Vehicle movement can be known from vehicle signals like speed and yaw rate, or be calculated from visual odometry. The most simple model of the environment is a flat world model, where the ground is flat and nothing exists above the ground. Several models could be used, e.g. a tunnel model can be used when driving in a tunnel.