METHOD FOR OPERATING A HEADLIGHT SYSTEM OF A MOTOR VEHICLE
20220396197 · 2022-12-15
Inventors
Cpc classification
B60Q1/143
PERFORMING OPERATIONS; TRANSPORTING
G01S7/4039
PHYSICS
B60Q1/06
PERFORMING OPERATIONS; TRANSPORTING
G01S13/87
PHYSICS
B60Q2300/30
PERFORMING OPERATIONS; TRANSPORTING
International classification
Abstract
A method for operating a headlight system of a motor vehicle, which during operation, illuminates an overall illumination area in the area surrounding the motor vehicle. At least two sensors are provided which detect objects in different associated sensor regions, the sensor regions of the sensors overlapping in an overlap region within the overall illumination area. The data of the sensors are fused to ascertain the existence of an object in the overlap region. Upon confirmation of the existence of the object, a zone in the overall illumination area in which the object is located is selectively excluded from illumination. A motor vehicle having a headlight system operated in this manner is also described.
Claims
1-10. (canceled)
11. A method for operating a headlight system of a motor vehicle, the method comprising the following steps: illuminating, by the headlight system, an overall illumination area in an area surrounding the motor vehicle; sensing, by each of at least two sensors, an associated sensor region in the area surrounding the motor vehicle, at least two of the sensors sensing sensor regions differing from each other, which overlap in an overlap region in the overall illumination area; fusing, upon detection of an object in the overlap region, sensor data of the sensors associated with the overlap region, and confirming existence of the object in the overall illumination area from the fused sensor data; in response to the confirmation of the existence of the object, selectively blocking out, by the headlight system, a zone in the overall illumination area in which the object is located.
12. The method as recited in claim 11, wherein the existence of the object in the overlap region is confirmed when both associated sensors detect the object in the overlap region and detected objects are in closer proximity to each other than a threshold value.
13. The method as recited in claim 11, wherein when an object in the overlap region is detected by only one of the associated sensors, a quality of the sensors is taken into account, and the existence of the object is confirmed when a sensor of the associated sensors having a higher quality detects the object.
14. The method as recited in claim 11, wherein upon detecting an object outside of the at least one overlap region, the existence of the object is confirmed.
15. A motor vehicle, comprising: a headlight system which is configured in such a way that during operation, it selectively illuminates an overall illumination area in an area surrounding the motor vehicle; at least two sensors, each of which during operation detects objects in an associated sensor region in the area surrounding the motor vehicle, the sensor regions of at least two of the sensors differing from each other and overlapping in an overlap region in the overall illumination area; wherein the motor vehicle is configured to: illuminate, using the headlight system, the overall illumination area in the area surrounding the motor vehicle, fuse, upon detection of an object in the overlap region, sensor data of the sensors associated with the overlap region, and confirm existence of the object in the overall illumination area from the fused sensor data, in response to the confirmation of the existence of the object, selectively block out from the illumination, by the headlight system, a zone in the overall illumination area in which the object is located.
16. The motor vehicle as recited in claim 15, wherein the motor vehicle has an optical camera as a first sensor of the at least two sensors, which during operation, senses a front-end section of the motor vehicle forward in a direction of travel as a first sensor region.
17. The motor vehicle as recited in claim 16, wherein the motor vehicle has at least one second sensor of the at least two sensors, which during operation, senses a second sensor region at a side in the direction of travel and extending laterally beyond the overall illumination area counter to the direction of travel, and forms an overlap region with at least one other sensor.
18. The motor vehicle as recited in claim 17, wherein at least one of the at least one second sensors is a radar sensor.
19. The motor vehicle as recited in claim 17, wherein the motor vehicle has a third sensor differing from the first sensor, which during operation, senses a front-end section of the motor vehicle forward in the direction of travel as a third sensor region.
20. The motor vehicle as recited in claim 19, wherein the third sensor is a radar sensor.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0044]
[0045]
[0046]
[0047]
DETAILED DESCRIPTION OF EXAMPLE EMBODIMENTS
[0048] A motor vehicle 1, as shown by way of example in
[0049] Motor vehicle 1 may be of any type. Purely by way of example, motor vehicle 1 in the exemplary embodiment of
[0050] To sense the area surrounding motor vehicle 1, motor vehicle 1 also has at least two sensors 8, each of which detects objects 10 in an associated sensor region 9, as can be gathered from
[0051] In the exemplary embodiments shown, motor vehicle 1 has an optical camera 12 as a first sensor 8a. During operation, first sensor 8a and thus camera 12 senses a front-end section of motor vehicle 1 forward in direction of travel 3 as associated sensor region 9a, which is also referred to hereinafter as first sensor region 9a. First sensor region 9a is also indicated with an associated angle a in
[0052]
[0053] The collected sensor data are evaluated by fusing the sensor data and evaluating them together in order to recognize objects 10 in overall illumination area 4 and to ascertain associated zones 5 for objects 10, as explained in the following with the aid of
[0054] A corresponding driving situation is illustrated in
[0055]
[0056] If the result in correspondence step 20 is negative, thus, object 10 was detected in one overlap region 11 and only by one of associated sensors 8, then in a following step 23, the quality of sensors 8 forming overlap region 11 is taken into account, this step 23 also being referred to hereinafter as quality-comparison step 23. In quality-comparison step 23, the qualities are compared in terms of whether sensor 8 having the higher quality has detected object 10. If this is the case, in a following step 24, the existence of object 10 is confirmed, this step 24 also being referred to hereinafter as single-sensor-object step 24. If sensor 8 having the lower quality has detected object 10, thus, the result of quality-comparison step 23 is negative, in a following step 25, the existence of object 10 is negated or rejected, or the existence of object 10 is assigned a low probability, this step 25 also being referred to hereinafter as rejection step 25.
[0057] As may also be gathered from
[0058] The results of the confirmation or the rejection of detected objects 10 are subsequently collected as consolidated or fused object data in a step 27, which hereinafter is also referred to as collection step 27. That is, the results of multi-sensor-object step 22, single-sensor-object step 24, rejection step 25 and single-sensor step 26 are fed to collection step 27. In a step 28, an associated zone 5 in overall illumination area 4 is then ascertained for respective confirmed object 10, this step 28 also being referred to hereinafter as zone-ascertainment step 28. The result of zone-ascertainment step 28 is supplied to headlight system 2, particularly control unit 16, which subsequently blocks out corresponding zones 5.
[0059] A reliable and precise recognition of objects 10 and ascertainment of associated zones 5, and consequently an improved block-out of objects 10 in overall illumination area 4 is thus accomplished.
[0060] It goes without saying that the method described is carried out continuously, especially in order to continuously block out objects 10 moving in overall illumination area 4.