Vehicle sensor with integrated radar and image sensors
11719808 · 2023-08-08
Assignee
Inventors
Cpc classification
B60T8/171
PERFORMING OPERATIONS; TRANSPORTING
B60T7/22
PERFORMING OPERATIONS; TRANSPORTING
B60T2201/022
PERFORMING OPERATIONS; TRANSPORTING
International classification
G01S13/86
PHYSICS
B60T7/22
PERFORMING OPERATIONS; TRANSPORTING
Abstract
A vehicular sensing system includes a sensor module disposed at the vehicle. The sensor module includes first and second radar sensors and a camera. A field of sensing of the first radar sensor is encompassed by a portion of a field of view of the camera and a field of sensing of the second radar sensor is encompassed by another portion of the field of view of the camera. Outputs of the radar sensors and the camera are communicated to a control. The control, responsive to processing of the outputs of the radar sensors, detects the presence of an object exterior the vehicle and within the field of sensing of at least one of the first and second radar sensors. The control, responsive to detection of an object via processing of the outputs of the radar sensors, processes the output of the camera to classify the detected object.
Claims
1. A vehicular sensing system, said vehicular sensing system comprising: a sensor module disposed at a vehicle equipped with said vehicular sensing system; wherein said sensor module comprises (i) a first radar sensor having a plurality of transmitting antennas and a plurality of receiving antennas, (ii) a second radar sensor having a plurality of transmitting antennas and a plurality of receiving antennas, and (iii) a camera; wherein said sensor module comprises a circuit board, and wherein said first and second radar sensors and said camera are disposed at said circuit board, and wherein said camera is disposed at said circuit board at a location between said first radar sensor and said second radar sensor; wherein physical orientation of the transmitting and receiving antennas of said second radar sensor at said circuit board relative to said camera is different from physical orientation of the transmitting and receiving antennas of said first radar sensor at said circuit board relative to said camera; wherein said first radar sensor has a first field of sensing exterior the vehicle, and wherein said second radar sensor has a second field of sensing exterior the vehicle, and wherein said camera has a field of view exterior the vehicle; wherein the first field of sensing of said first radar sensor is encompassed by a portion of the field of view of said camera, and wherein the second field of sensing of said second radar sensor is encompassed by another portion of the field of view of said camera; a control, wherein outputs of said first and second radar sensors and said camera are communicated to said control; wherein said control comprises a processor that processes the outputs of said first and second radar sensors and said camera that are communicated to said control; wherein said control, responsive to processing of the outputs of said first and second radar sensors, detects the presence of an object exterior the vehicle and within the field of sensing of at least one of said first radar sensor and said second radar sensor; and wherein said control, responsive to detection of the presence of an object via processing of the outputs of said first and second radar sensors, processes the output of said camera to classify the detected object.
2. The vehicular sensing system of claim 1, wherein said first and second radar sensors are positioned such that their horizontal and vertical azimuths are intersecting.
3. The vehicular sensing system of claim 1, wherein said first and second radar sensors and said camera are disposed in a common housing.
4. The vehicular sensing system of claim 1, wherein radar signals transmitted by said plurality of transmitting antennas of said first radar sensor are received by said plurality of receiving antennas of said second radar sensor, and wherein radar signals transmitted by said plurality of transmitting antennas of said second radar sensor are received by said plurality of receiving antennas of said first radar sensor.
5. The vehicular sensing system of claim 1, wherein said sensor module is disposed at a rear portion of the vehicle, and wherein said camera views rearward of the vehicle.
6. The vehicular sensing system of claim 5, wherein, during a reversing maneuver of the vehicle, a display device disposed in a cabin of the vehicle and viewable by a driver of the vehicle displays video images derived from the output of said camera.
7. The vehicular sensing system of claim 1, wherein a display device is disposed in a cabin of the vehicle and is viewable by a driver of the vehicle, and wherein the display device is operable to display video images derived from the output of said camera.
8. The vehicular sensing system of claim 1, wherein the transmitting antennas and the receiving antennas of said sensor module are operable to function as virtual antennas.
9. The vehicular sensing system of claim 1, wherein the transmitting antennas and the receiving antennas of said sensor module are operable to function as at least four virtual antennas.
10. The vehicular sensing system of claim 1, wherein said sensor module comprises part of a driving assist system of the vehicle.
11. The vehicular sensing system of claim 10, wherein said vehicular sensing system provides object detection for at least one selected from the group consisting of (i) automated parking, (ii) blind spot detection, (iii) cross traffic alert, (iv) lane change and merge aid, (v) automatic emergency braking, (vi) pedestrian detection, (vii) turn assist, (viii) terrain mapping and (ix) intersection collision mitigation.
12. The vehicular sensing system of claim 1, wherein a first physical orientation of the transmitting and receiving antennas of said first radar sensor relative to a first axis of said first radar sensor is rotationally offset different than a second physical orientation of the transmitting and receiving antennas of said second radar sensor relative to a second axis of said second radar sensor.
13. The vehicular sensing system of claim 12, wherein the first physical orientation of the transmitting and receiving antennas of said first radar sensor is rotationally offset clockwise about the first axis relative to said camera, and wherein the second physical orientation of the transmitting and receiving antennas of said second radar sensor is rotationally offset counter-clockwise about the second axis relative to said camera.
14. The vehicular sensing system of claim 13, wherein the first axis is parallel to the second axis.
15. The vehicular sensing system of claim 12, wherein the first physical orientation of the transmitting and receiving antennas of said first radar sensor is rotationally offset by a selected angle about the first axis relative to said camera, and wherein the second physical orientation of the transmitting and receiving antennas of said second radar sensor is rotationally offset by the selected angle about the second axis relative to said camera.
16. The vehicular sensing system of claim 15, wherein the selected angle is less than or equal to 45 degrees.
17. The vehicular sensing system of claim 15, wherein the selected angle is 35 degrees.
18. The vehicular sensing system of claim 12, wherein the first physical orientation of the transmitting and receiving antennas of said first radar sensor is rotationally offset 90 degrees about the first axis relative the second physical orientation of the transmitting and receiving antennas of said second radar sensor.
19. A vehicular sensing system, said vehicular sensing system comprising: a sensor module disposed at a vehicle equipped with said vehicular sensing system; wherein said sensor module comprises (i) a first radar sensor having a plurality of transmitting antennas and a plurality of receiving antennas, (ii) a second radar sensor having a plurality of transmitting antennas and a plurality of receiving antennas, and (iii) a camera; wherein said sensor module comprises a circuit board, and wherein said first and second radar sensors and said camera are disposed at said circuit board, and wherein said camera is disposed at said circuit board at a location between said first radar sensor and said second radar sensor; wherein said first radar sensor has a first field of sensing exterior the vehicle, and wherein said second radar sensor has a second field of sensing exterior the vehicle, and wherein said camera has a field of view exterior the vehicle; wherein the first field of sensing of said first radar sensor is encompassed by a portion of the field of view of said camera, and wherein the second field of sensing of said second radar sensor is encompassed by another portion of the field of view of said camera; wherein said sensor module is disposed at a rear portion of the vehicle, and wherein said camera views rearward of the vehicle; a control, wherein outputs of said first and second radar sensors and said camera are communicated to said control; wherein said control comprises a processor that processes the outputs of said first and second radar sensors and said camera that are communicated to said control; wherein said control, responsive to processing of the outputs of said first and second radar sensors, detects the presence of an object exterior the vehicle and within the field of sensing of at least one of said first radar sensor and said second radar sensor; wherein said control, responsive to detection of the presence of an object via processing of the outputs of said first and second radar sensors, processes the output of said camera to classify the detected object; and wherein, during a reversing maneuver of the vehicle, a display device disposed in a cabin of the vehicle and viewable by a driver of the vehicle displays video images derived from the output of said camera.
20. The vehicular sensing system of claim 19, wherein said first and second radar sensors are positioned such that their horizontal and vertical azimuths are intersecting.
21. The vehicular sensing system of claim 19, wherein radar signals transmitted by said plurality of transmitting antennas of said first radar sensor are received by said plurality of receiving antennas of said second radar sensor, and wherein radar signals transmitted by said plurality of transmitting antennas of said second radar sensor are received by said plurality of receiving antennas of said first radar sensor.
22. A vehicular sensing system, said vehicular sensing system comprising: a sensor module disposed at a vehicle equipped with said vehicular sensing system; wherein said sensor module comprises (i) a first radar sensor having a plurality of transmitting antennas and a plurality of receiving antennas, (ii) a second radar sensor having a plurality of transmitting antennas and a plurality of receiving antennas, and (iii) a camera; wherein said camera is disposed between said first radar sensor and said second radar sensor; wherein physical orientation of the transmitting and receiving antennas of said second radar sensor relative to said camera is different from physical orientation of the transmitting and receiving antennas of said first radar sensor relative to said camera; wherein said first radar sensor has a first field of sensing exterior and at least forward of the vehicle, and wherein said second radar sensor has a second field of sensing exterior and at least forward of the vehicle, and wherein said camera has a field of view exterior and at least forward of the vehicle; wherein the first field of sensing of said first radar sensor is encompassed by a portion of the field of view of said camera, and wherein the second field of sensing of said second radar sensor is encompassed by another portion of the field of view of said camera; a control, wherein outputs of said first and second radar sensors and said camera are communicated to said control; wherein said control comprises a processor that processes the outputs of said first and second radar sensors and said camera that are communicated to said control; wherein said control, responsive to processing of the outputs of said first and second radar sensors, detects the presence of an object exterior the vehicle and within the field of sensing of at least one of said first radar sensor and said second radar sensor; and wherein said control, via processing of the outputs of said first and second radar sensors, provides object detection for an automatic emergency braking system of the vehicle.
23. The vehicular sensing system of claim 22, wherein said control, responsive to detection of the presence of an object via processing of the outputs of said first and second radar sensors, processes the output of said camera to classify the detected object.
24. The vehicular sensing system of claim 22, wherein said first and second radar sensors are positioned such that their horizontal and vertical azimuths are intersecting.
25. The vehicular sensing system of claim 22, wherein said first and second radar sensors and said camera are disposed in a common housing.
26. The vehicular sensing system of claim 22, wherein said vehicular sensing system provides object detection for at least one selected from the group consisting of (i) automated parking, (ii) blind spot detection, (iii) cross traffic alert, (iv) lane change and merge aid, (v) automatic emergency braking, (vi) pedestrian detection, (vii) turn assist, (viii) terrain mapping and (ix) intersection collision mitigation.
27. The vehicular sensing system of claim 22, wherein a first physical orientation of the transmitting and receiving antennas of said first radar sensor relative to a first axis of said first radar sensor is rotationally offset different than a second physical orientation of the transmitting and receiving antennas of said second radar sensor relative to a second axis of said second radar sensor.
28. The vehicular sensing system of claim 27, wherein the first physical orientation of the transmitting and receiving antennas of said first radar sensor is rotationally offset clockwise about the first axis relative to said camera, and wherein the second physical orientation of the transmitting and receiving antennas of said second radar sensor is rotationally offset counter-clockwise about the second axis relative to said camera.
29. The vehicular sensing system of claim 28, wherein the first axis is parallel to the second axis.
30. The vehicular sensing system of claim 27, wherein the first physical orientation of the transmitting and receiving antennas of said first radar sensor is rotationally offset by a selected angle about the first axis relative to said camera, and wherein the second physical orientation of the transmitting and receiving antennas of said second radar sensor is rotationally offset by the selected angle about the second axis relative to said camera.
31. The vehicular sensing system of claim 30, wherein the selected angle is less than or equal to 45 degrees.
32. The vehicular sensing system of claim 27, wherein the first physical orientation of the transmitting and receiving antennas of said first radar sensor is rotationally offset 90 degrees about the first axis relative the second physical orientation of the transmitting and receiving antennas of said second radar sensor.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1)
(2)
(3)
(4)
(5)
(6)
(7)
(8)
DESCRIPTION OF THE PREFERRED EMBODIMENTS
(9) A vehicle sensing system, such as a driver or driving assist system, object detection system, parking assist system and/or alert system, operates to capture sensing data exterior of the vehicle and may process the captured data to detect objects at or near the vehicle and in the predicted path of the vehicle, such as to assist a driver of the vehicle in maneuvering the vehicle in a forward or rearward direction or to assist the driver in parking the vehicle in a parking space. The system includes a processor that is operable to receive sensing data from multiple sensors and to provide an output to a control that, responsive to the output, generates an alert or controls an accessory or system of the vehicle, or highlights or overlays an alert on a display screen (that may be displaying video images captured by a single rearward viewing camera or multiple cameras providing forward, side or 360 degree surround views of the area surrounding the vehicle during a reversing or low speed maneuver of the vehicle).
(10) Referring now to the drawings and the illustrative embodiments depicted therein, a vehicle 10 includes an driver assistance system or sensing system 12 that includes at least one radar sensor unit, such as a forward facing radar sensor unit 14 (and the system may optionally include multiple exterior facing sensors, such as cameras or other sensors, such as a rearward facing sensor at the rear of the vehicle, and a sideward/rearward facing sensor at respective sides of the vehicle), which sense regions exterior of the vehicle. The sensing system 12 includes a control or electronic control unit (ECU) or processor that is operable to process data captured by the sensor or sensors and may detect objects or the like. The data transfer or signal communication from the sensor to the ECU may comprise any suitable data or communication link, such as a vehicle network bus or the like of the equipped vehicle.
(11) Some automotive radars use MIMO (Multiple Input Multiple Output) techniques to create an effective virtual antenna aperture, which is significantly larger than the real antenna aperture, and delivers much better angular resolution than conventional radars, such as, for example, conventional scanning radars. MIMO techniques may be used to create virtual antenna apertures, not only from linear arrays of real antennas but also from two dimensional arrays of real antennas. For example, the antenna array of a MIMO sensor may comprise two transmitting antennas and two or more receiving antennas, arranged either in a one-dimensional array or as in a two-dimensional array. Thus, the antenna array may consist of four antennas (arranged as a 2×2 array MIMO virtual antenna) or any other combination of transmitting and receiving antennas, whose product is the number of virtual antennas, such as, for example, 16 virtual antennas (4×4 array) or more or less than 16 virtual antennas.
(12) For vision systems having an exterior viewing camera, the challenge is that the OEM system level requirements may specify that objects need to be detected and classified between 0 m . . . 18 m with a resolution in the 10 cm range. In order to ‘see’ enough of the object in the close range (0 . . . 2 m), a camera lens with about a 185 degree opening angle is needed producing a fisheye image. The large opening angle of the lens in combination with an imager with limited resolution (such as around 1 MP, which are spread in an angle to cover the desired wide field of view) causes a very poor resolution in the long range. For example, at 10 m range the pixel resolution is around 1.2 m, meaning that distance measurement with enough precision beyond 10 m is difficult to nearly impossible.
(13) For a radar sensing system with a 16 virtual receiver configuration, the angular resolution is approximately 10 degrees. The spot size of a slice of the radar's field of view would be approximated by the equation (angular resolution×range)/57. Thus, at one meter, the spot slice is about 17 cm. As the range is extended, the spot size would increase.
(14) If a radar sensor with horizontal slices and another radar sensor with vertical slices are utilized to cover an FOV, the intersection of their slices one from each sensor permits a 17 square cm spot in 2-D for fine resolution over the common FOV. The BSD (blind spot detection) corner sensors would allow for conversion of the spot to a cube for three dimensional (3D) and terrain monitoring.
(15) Corner radar sensors may be used for detection and cameras may be used for classification. For example, the radar sensors mounted at the corners of the vehicle are used for the object detection and the rearward viewing camera may be used for classification (and optionally for display of video images derived from and representative of image data captured by the rearward viewing camera). The radar sensors can be used for measurement of object height and object location. A higher accuracy is achieved in the FOV where the radar sensors overlap. However, a radar blindspot may occur at the center area of the vehicle (see
(16) The present invention provides an integrated radar sensor and camera that are integrated into the same or common housing (see
(17) For example, the radar and camera data may be fused for an automatic emergency braking (AEB) system of the vehicle. The machine vision/image processing of the camera does not require object detection anymore, therefore is of less risk for smart camera implementations (less processing performance required). If the camera and radar are packaged together, there is less integration effort for the OEM, also wiring is minimized.
(18) Optionally, the system may integrate radar sensor(s) in stereo three dimensional (3D) configuration and a camera into a common housing. The radar sensors are positioned such that their horizontal and vertical azimuths are intersecting. It is envisioned that the radar sensors have common horizontal and vertical fields of view or fields of sensing and common azimuth resolution. Each radar sensor is used for object detection and object localization, with a vertical opening angle of about 150 degrees, a range of about 30+ m, and a range resolution of 7.5 cm. The radar sensor allows measuring of the height of objects and creation of a terrain map. The pitch of the radar sensor mounting (see
(19) It is envisioned that the sensors' horizontal planes are rotated greater than about 90 degrees relative to each other, providing intersecting beams in range, velocity, azimuth and elevation for all detections, creating radar data cubes of known size for all volumes in 3D space within the combined field of view of at least three sensors on the vehicle. Camera, radar and machine vision/image processing is used for object classification (such as, for example, pedestrians, vehicles, bicycles, and/or the like). The captured radar data and camera/image data is fused for automatic emergency braking (AEB) applications.
(20) Optionally, the radar sensors may operate collaboratively, where the transmitted signals from each transmit antenna of radar sensor #1 are received and processed by all receive antenna of both radar sensor #1 and radar sensor #2, thereby increasing effective aperture, accuracy and resolution of the system.
(21) The machine vision/image processing of the camera-captured image data does not require object detection, and thus is of less risk for smart camera implementations (less processing performance required). If the camera and radar are packaged together, there is less integration effort for the OEM, while also minimizing or reducing wiring requirements.
(22) As shown in
(23) The system may utilize sensors, such as radar or lidar sensors or the like. The sensing system may utilize aspects of the systems described in U.S. Pat. Nos. 9,753,121; 9,689,967; 9,599,702; 9,575,160; 9,146,898; 9,036,026; 8,027,029; 8,013,780; 6,825,455; 7,053,357; 7,408,627; 7,405,812; 7,379,163; 7,379,100; 7,375,803; 7,352,454; 7,340,077; 7,321,111; 7,310,431; 7,283,213; 7,212,663; 7,203,356; 7,176,438; 7,157,685; 6,919,549; 6,906,793; 6,876,775; 6,710,770; 6,690,354; 6,678,039; 6,674,895 and/or 6,587,186, and/or International Publication No. WO 2011/090484 and/or U.S. Publication Nos. US-2017-0222311 and/or US-2010-0245066, and/or U.S. patent applications, Ser. No. 15/647,339, filed Jul. 12, 2017, now U.S. Pat. No. 10,239,446, Ser. No. 15/619,627, filed Jun. 12, 2017, now U.S. Pat. No. 10,768,298, Ser. No. 15/584,265, filed May 2, 2017, now U.S. Pat. No. 10,534,081, Ser. No. 15/467,247, filed Mar. 23, 2017, now U.S. Pat. No. 10,571,562, Ser. No. 15/446,220, filed Mar. 1, 2017, now U.S. Pat. No. 10,863,335, and/or Ser. No. 15/675,919, filed Aug. 14, 2017, now U.S. Pat. No. 10,641,867, and/or International PCT Application No. PCT/IB2017/054120, filed Jul. 7, 2017, and published on Jan. 11, 2018 as International Publication No. WO 2018007995, which are hereby incorporated herein by reference in their entireties.
(24) Changes and modifications in the specifically described embodiments can be carried out without departing from the principles of the invention, which is intended to be limited only by the scope of the appended claims, as interpreted according to the principles of patent law including the doctrine of equivalents.