Method for the animated representation of an object perception and of a driving intention of an assistance system of a vehicle, assistance system, computer program, and computer-readable (storage) medium
12559125 ยท 2026-02-24
Assignee
Inventors
Cpc classification
B60W50/14
PERFORMING OPERATIONS; TRANSPORTING
B60K2360/161
PERFORMING OPERATIONS; TRANSPORTING
B60K2360/179
PERFORMING OPERATIONS; TRANSPORTING
B60K35/29
PERFORMING OPERATIONS; TRANSPORTING
B62D15/029
PERFORMING OPERATIONS; TRANSPORTING
B60K2360/1868
PERFORMING OPERATIONS; TRANSPORTING
B60W2540/215
PERFORMING OPERATIONS; TRANSPORTING
B60W2555/60
PERFORMING OPERATIONS; TRANSPORTING
B60K35/10
PERFORMING OPERATIONS; TRANSPORTING
B60K35/28
PERFORMING OPERATIONS; TRANSPORTING
International classification
B60W50/14
PERFORMING OPERATIONS; TRANSPORTING
B60K35/28
PERFORMING OPERATIONS; TRANSPORTING
Abstract
Systems, methods, and apparatuses are provided for animated representation of at least one object perception and of a driving intention of an assistance system for at least partially automated maneuvering of a vehicle. An object is perceived within an environment of the vehicle using the assistance system. The perceived object influences a control strategy of the at least partially automated maneuvering of the assistance system. The perceived object is represented. The vehicle is represented by an avatar. A first animation is displayed of an interaction between the avatar and the perceived object. The avatar is animated to mimic a first action based on the perceived object. A second animation is displayed representing the driving intention of the assistance system. The avatar is animated to mimic a second action based on a plan of the control strategy influenced by the perceived object.
Claims
1. A method for representing at least one object perception and a driving intention of an assistance system for at least partially automated maneuvering of a vehicle, comprising: perceiving an object within an environment of the vehicle using the assistance system, wherein; the perceived object influences a control strategy of the at least partially automated maneuvering of the assistance system; representing the perceived object as an animated representation; representing the vehicle by an avatar; displaying a first animation of an interaction between the avatar and the perceived object, wherein: the avatar is animated to mimic performing a first action associated with the avatar based on perceiving, by the assistance system, the perceived object; and displaying a second animation representing the driving intention of the assistance system, wherein: the avatar is animated to mimic performing, within the second animation, a second action associated with the avatar directly on the animated representation of the perceived object based on a plan of the control strategy being influenced by the perceived object, wherein the second animation conveys the plan of the assistance system to a driver of the vehicle.
2. The method according to claim 1, wherein; an animation parameter is determined as a function of the perceived object and/or as a function of the driving intention of the assistance system, and the animation parameter influences the first and/or second animation.
3. The method according to claim 2, wherein: the animation parameter describes an urgency of an implementation of the control strategy, and the first and/or second animation occurs as a function of the urgency.
4. The method of claim 3, wherein: the animation parameter corresponds to a speed of the first and/or second animation; and the first and/or second animation occurs as the function of the urgency based on influencing the speed of the first and/or second animation.
5. The method according to claim 1, wherein; the first and/or second animation occurs such that the perceived object or the driving intention of the assistance system is displayed within peripheral vision of the driver of the vehicle.
6. The method of claim 1, wherein: the avatar comprises a non-motor vehicle representation of the vehicle.
7. The method of claim 1, wherein: based on the avatar comprising a dragon, the first action comprises the avatar spitting fire in a direction of the animated representation within the first animation and the second action comprises the avatar devouring the animated representation within the second animation; or based on the avatar comprising a riding cowboy, the first action comprises the avatar whirling a lasso within the first animation and the second action comprises the avatar ensnaring the animated representation with the lasso within the second animation.
8. The method of claim 1, wherein: the avatar representing the vehicle varies based on the perceived object and/or the driving intention of the assistance system.
9. An assistance system for a vehicle, comprising: a display configured to represent the perceived object and the driving intention of the assistance system in an animated manner in accordance with a method according to claim 1.
10. A non-transitory computer-readable medium comprising instructions operable, when executed by one or more computing systems, to: perceive an object within an environment of a vehicle using an assistance system, wherein; the perceived object influences a control strategy of an at least partially automated maneuvering of the assistance system; represent the perceived object as an animated representation; represent the vehicle by an avatar; display a first animation of an interaction between the avatar and the perceived object, wherein: the avatar is animated to mimic performing a first action associated with the avatar based on perceiving, by the assistance system, the perceived object; and display a second animation representing a driving intention of the assistance system, wherein: the avatar is animated to mimic performing, within the second animation, a second action associated with the avatar directly on the animated representation of the perceived object based on a plan of the control strategy being influenced by the perceived object, wherein the second animation conveys the plan of the assistance system to a driver of the vehicle.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1)
(2)
(3)
DETAILED DESCRIPTION OF THE DRAWINGS
(4) In the figures, identical or functionally identical elements are provided with the same reference symbols.
(5)
(6) The assistance system may, for instance, take the form of an adaptive cruise control system with traffic-sign recognition, proximity control and/or such like. The display device 3 represents the object perception of the assistance system schematically. Within the scope of the object perception, the assistance system captures the stop-sign 5. The stop-sign 5 in the environment 4 of the vehicle 1 is represented in the display device 3 by the virtual stop-sign 5. Moreover, the stopping-line 6 in the environment 4 of the vehicle 1 is represented in the display device 3 by the virtual stopping-line 6. The vehicle 1 is represented by an avatar 7. In the example shown in
(7) In the example shown in
(8) Under certain circumstances the driver is accordingly still less aware that the control strategy of the at least partially automated maneuvering by the assistance systemthat is to say, in particular, the longitudinal control of the vehicle 1 by the assistance systemis being influenced. Accordingly, under certain circumstances the driver of the vehicle 1 is not cognizant that the stop-sign 5 has been captured and that the assistance system is thereupon changing the control strategy. In other words, under certain circumstances the driver is accordingly unaware that the assistance system of the vehicle 1 intends to stop at the stopping-line 6.
(9) The fact that the stop-sign 5 is captured by the assistance system within the scope of the object perception can be made clear to the driver of the vehicle 1 within the scope of the animating of the at least one object perception by virtue of at least one additional action that the avatarrepresented here by spitting of fire 9mimics. Alternatively, the additional action that the avatar mimics may be a devouring of the virtual stop-sign 5.
(10) In summary, using the method according to the present subject matter for animated representation of the at least one object perception and of the driving intention of the assistance system it can accordingly be ensured that the driver perceives relevant information pertaining to the assistance system better, and as a result is given an improved understanding of the system.
(11)
(12) The vehicle 1 is traveling in the environment 4 on a freeway 10 having a roadway that has three lanes. The vehicle 1 is traveling in the left lane 14 of the roadway. Over and above this, a lane ends panel 11 marks the end of the left lane 14.
(13) Further road-users 12 are located in the middle lane. When the assistance system, which in this example controls at least a lateral and longitudinal guidance of the vehicle 1, has been activated, in the example represented here the driver of the vehicle 1 is not aware of whether the lane ends panel 11 and the end of the left lane 14 were captured correctly by the assistance system. This can be presented to the driver of the vehicle 1 by the display device 3.
(14) In the example from
(15) However, under certain circumstances at this point in time the driver of the vehicle 1 is not aware of the consequences that arise for the control strategy, or the driving intention, of the assistance system by reason of the lane ends panel 11. In other words, it may accordingly be the case that the particular driving intention of the assistance system is not clear to the driver. Accordingly, the assistance system might, for instance, decelerate the vehicle 1 considerably in the left lane 14 and cut in behind the two further road-users 12. Similarly, it is possible that the assistance system has the driving intention to accelerate the vehicle 1, in order subsequently to cut in between the two further road-users 12.
(16) In the example from
(17) An animation parameter can, for instance, be determined as a function of how strongly the vehicle 1 has to be accelerated in order to cut in between the two further road-users 12. Such an animation parameter can describe at least an urgency of an implementation of the control strategythat is to say, for instance, of the acceleratingand can influence a speed of animation. As a result, an urgency can be signaled to the driver of the vehicle 1.
(18)