Identification of attention region for enhancement of sensor-based detection in a vehicle
10984534 · 2021-04-20
Assignee
Inventors
- Wei Tong (Troy, MI, US)
- Shuqing Zeng (Sterling Heights, MI)
- Upali P. Mudalige (Oakland Township, MI, US)
Cpc classification
G06V30/2504
PHYSICS
B60R11/04
PERFORMING OPERATIONS; TRANSPORTING
G06V20/70
PHYSICS
G06F18/254
PHYSICS
G06V20/58
PHYSICS
G06V10/25
PHYSICS
G06V10/809
PHYSICS
International classification
B60R11/04
PERFORMING OPERATIONS; TRANSPORTING
Abstract
Systems and methods to identify an attention region in sensor-based detection involve obtaining a detection result that indicates one or more detection areas where one or more objects of interest are detected. The detection result is based on using a first detection algorithm. The method includes obtaining a reference detection result that indicates one or more reference detection areas where one or more objects of interest are detected. The reference detection result is based on using a second detection algorithm. The method also includes identifying the attention region as one of the one or more reference detection areas without a corresponding one or more detection areas. The first detection algorithm is used to perform detection in the attention region.
Claims
1. A method of identifying an attention region in sensor-based detection, the method comprising: obtaining, by processing circuitry, a detection result that indicates one or more detection areas where one or more objects of interest are detected by a sensor, wherein the detection result is based on using a first detection algorithm; obtaining, by the processing circuitry, a reference detection result that indicates one or more reference detection areas where one or more objects of interest are detected by the sensor or by a different sensor, wherein the reference detection result is based on using a second detection algorithm; identifying, by the processing circuitry, the attention region as one of the one or more reference detection areas without a corresponding one or more detection areas; and the processing circuitry using the first detection algorithm to perform detection in the attention region.
2. The method according to claim 1, further comprising obtaining data from a first sensor at a first resolution.
3. The method according to claim 2, further comprising reducing a resolution of the data to a second resolution.
4. The method according to claim 3, wherein the obtaining the detection result includes using the first detection algorithm on the data at the second resolution.
5. The method according to claim 3, wherein the obtaining the reference detection result includes using the second detection algorithm on the data at the second resolution.
6. The method according to claim 2, wherein the obtaining the reference detection result includes using the second detection algorithm on other data from another sensor.
7. The method according to claim 6, wherein the obtaining the detection result and the obtaining the reference detection result includes the first sensor and the another sensor having a common field of view.
8. The method according to claim 2, wherein the obtaining the data from the first sensor includes obtaining the data from a camera.
9. The method according to claim 8, wherein the obtaining the detection result includes obtaining one or more bounding boxes that surround a detected object.
10. The method according to claim 8, wherein the obtaining the reference detection result includes obtaining one or more classification areas resulting from segmentation, each classification area indicating an object classification in an area of a field of view of the camera.
11. A system to identify an attention region in sensor-based detection, the system comprising: a sensor configured to obtain data; and a processor configured to obtain a detection result that indicates one or more detection areas where one or more objects of interest are detected from the data, wherein the detection result is based on using a first detection algorithm, to obtain a reference detection result that indicates one or more reference detection areas where one or more objects of interest are detected, wherein the reference detection result is based on using a second detection algorithm, to identify the attention region as one of the one or more reference detection areas without a corresponding one or more detection areas, and to use the first detection algorithm to perform detection in the attention region.
12. The system according to claim 11, wherein the data from the sensor is at a first resolution.
13. The system according to claim 12, wherein the processor is further configured to reduce a resolution of the data to a second resolution.
14. The system according to claim 13, wherein the processor obtains the detection result using the first detection algorithm on the data at the second resolution.
15. The system according to claim 13, wherein the processor obtains the reference detection result using the second detection algorithm on the data at the second resolution.
16. The system according to claim 12, wherein the processor obtains the reference detection result using the second detection algorithm on other data from another sensor.
17. The system according to claim 16, wherein the sensor and the other sensor have a common field of view.
18. The system according to claim 16, wherein the sensor and the other sensor are in a vehicle.
19. The system according to claim 12, wherein the first sensor is a camera.
20. The system according to claim 19, wherein the detection result includes one or more bounding boxes that surround a detected object, and the reference detection result includes one or more classification areas resulting from segmentation, each classification area indicating an object classification in an area of a field of view of the camera.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) Other features, advantages and details appear, by way of example only, in the following detailed description, the detailed description referring to the drawings in which:
(2)
(3)
(4)
DETAILED DESCRIPTION
(5) The following description is merely exemplary in nature and is not intended to limit the present disclosure, its application or uses. It should be understood that throughout the drawings, corresponding reference numerals indicate like or corresponding parts and features.
(6) As previously noted, data obtained with sensors may be processed to obtain information such as information about objects around a vehicle. When high resolution sensor data is processed naively (i.e., without any additional knowledge), the resulting detection may not only be inaccurate but also consumes resources without yielding useful results. Embodiments of the systems and methods detailed herein relate to the identification of an attention region for detection enhancement in a vehicle. Attention region is a term used to refer to a region that warrants further analysis. Specifically, low resolution data is used to identify the attention region, which is then analyzed further using the high resolution data. In this way, resources are expended for high resolution processing only in regions that are known to be of interest.
(7) In accordance with an exemplary embodiment,
(8) The controller 120 may also implement or communicate with other controllers that implement autonomous or semi-autonomous vehicle functions (e.g., autonomous braking, collision avoidance, adaptive cruise control, autonomous driving) based on information obtained by the camera 110 or other sensors 130. The information may include the relative position of objects 140 such as the pedestrian 150 shown in
(9)
(10) At block 230, obtaining a reference detection result may be performed in different ways. In the case of the low-resolution image (at block 220), the reference detection result may be a pixel-level classification referred to as segmentation. That is, the low-resolution image may be partitioned (i.e., segmented) and each segment (i.e., pixel region) may be classified based on the values of the pixels in the segment. The classification corresponds with objects of interest in the image, and the segmentation results in a set of pixel regions S={s.sub.i}, i=1, . . . , m and m is the total number of pixel regions. Thus, each s.sub.i provides the classification associated with a pixel region. According to an alternate embodiment, the reference detection result may be a detection result obtained, at block 230, using a different sensor 130 (e.g., radar system 133, lidar system 135). Thus, the reference detection result may be based on using a different type of detection (e.g., segmentation) on the reduced-resolution data (at block 220) from the same sensor 130 or on using a different sensor 130. When the reference detection result is obtained using a different sensor 130, the fields of view of the first sensor and the different sensor 130 must at least overlap in order to facilitate identification of an attention region.
(11) At block 240, obtaining a detection result refers to using the reduced-resolution data (at block 220) to perform object detection. This refers to implementing the typical object detection algorithm associated with the sensor 130. In the exemplary case of the sensor 130 being the camera 110, the result of the object detection (at block 240) may be a set of bounding boxes B={b.sub.j}, j=1, . . . , n, where n is the total number of bounding boxes. Each bounding box b.sub.j is a shape (e.g., rectangle) that bounds a region of the low-resolution image and indicates an object class for the region. In embodiments with a different sensor 130, obtaining the detection (at block 240) will be based on the typical detection modality for the sensor 130. For example, processing including beamforming may be used as part of the detection process for data obtained with a radar system 133.
(12) At block 250, determining a difference between the reference detection result (at block 230) and the detection result (at block 240) may result in identifying at least one attention area. The steps involved in identifying an attention area are detailed with reference to
(13)
(14) The processes at block 320 and 330 are performed for each remaining s.sub.i in S after block 310. The processes at blocks 320 and 330 may only be performed for remaining s.sub.i that have more than a threshold number of pixels. At block 320, mapping to the original resolution data refers to mapping the pixel region s.sub.i to s.sub.i′ in the original image (at block 210,
(15) While the above disclosure has been described with reference to exemplary embodiments, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted for elements thereof without departing from its scope. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the disclosure without departing from the essential scope thereof. Therefore, it is intended that the present disclosure not be limited to the particular embodiments disclosed, but will include all embodiments falling within the scope thereof