System and method of unlocking a vehicle based on visual domain data and wireless domain data
11613231 · 2023-03-28
Assignee
Inventors
- Bo Yu (Troy, MI, US)
- Prachi Joshi (Sterling Heights, MI, US)
- Fan Bai (Ann Arbor, MI)
- Madhu Max Seenisamy (Birmingham, MI, US)
- Ramesh Ramiah (Troy, MI, US)
- Kamran Ali (Troy, MI, US)
Cpc classification
G06V20/59
PHYSICS
B60R25/2081
PERFORMING OPERATIONS; TRANSPORTING
H04N23/90
ELECTRICITY
B60R25/245
PERFORMING OPERATIONS; TRANSPORTING
G06V20/56
PHYSICS
International classification
B60R25/20
PERFORMING OPERATIONS; TRANSPORTING
B60R25/24
PERFORMING OPERATIONS; TRANSPORTING
B60R25/30
PERFORMING OPERATIONS; TRANSPORTING
G06V10/98
PHYSICS
G06V20/56
PHYSICS
G06V20/59
PHYSICS
Abstract
A method of unlocking a vehicle of a user having a handheld device based on visual domain and radio frequency (RF) domain technologies is provided. The method comprises activating an on-board apparatus of the vehicle when the handheld device of the user is within a threshold distance from the vehicle. The method further comprises comparing a visual domain position of the user based on visual domain data with an RF domain position of the user based on RF domain data by a first equation, |d.sub.visual.sup.t.sup.
Claims
1. A method of unlocking a vehicle of a user having a handheld device based on visual domain and radio frequency (RF) domain technologies, the method comprising: activating an on-board apparatus of the vehicle when the handheld device of the user is within a threshold distance from the vehicle; comparing a visual domain position of the user based on visual domain data with an RF domain position of the user based on RF domain data by a
first equation,
|d.sub.visual.sup.t.sup.
and a second equation,
|φ.sub.visual.sup.t.sup.
2. The method of claim 1 wherein the step of activating comprises: monitoring a location of the vehicle and a position of a handheld device of a user based on an intentional detection module; and detecting whether the handheld device is within the threshold distance from the vehicle.
3. The method of claim 1 wherein the step of comparing comprises: verifying a facial identity of the user based on visual domain data of the user; estimating the visual domain position of the user based on visual domain data of the user; determining an RF domain position of the vehicle based on RF domain data of the vehicle; and determining the RF domain position of the user based on RF domain data of the handheld device.
4. The method of claim 3 wherein the step of comparing further comprises: detecting a non-user having a non-user device based on visual domain data of the non-user and RF domain data of the non-user device to define a visual domain position and an RF domain position of the non-user; comparing the visual domain position and the RF domain position of the user with the visual domain position and the RF domain position of the non-user by a third equation,
|d.sub.visual.sup.t.sup.
a fourth equation,
φ.sub.visual.sup.t.sup.
a fifth equation,
|d.sub.RF.sup.t.sup.
a sixth equation,
|φ.sub.RF.sup.t.sup.
5. The method of claim 1 further comprising performing a sequence of actions within the vehicle based on personalized settings of the user after unlocking the vehicle.
6. The method of claim 1 further comprising; detecting the facial identity and the visual domain position of the user based on visual domain data of the user when the user is in the vehicle; and verifying the facial identity of the user based on visual domain data of the user when the user is in the vehicle.
7. The method of claim 1 further comprising: detecting a companion-user having a companion-user device based on visual domain data of the companion-user and RF domain data of the companion-user device to define a visual domain position and an RF domain position of the companion-user; and performing a sequence of actions within the vehicle based on personalized settings of the companion-user after unlocking the vehicle.
8. The method of claim 3 wherein verifying the facial identity of the user comprises: assessing image data quality of the facial identity of the user based on visual domain data of the user; determining a confidence threshold based on the image quality of the facial identity of the user; processing the visual domain data of the user with a facial recognition module to define a confidence score of the facial identity of the user; and confirming the facial identity of the user with an authentication module when the confidence score is greater than the confidence threshold.
9. A system for unlocking a vehicle of a user having a handheld device based on visual domain and radio frequency (RF) domain technologies, the system comprising: an onboard apparatus of the vehicle, the onboard apparatus comprising: an electronic control unit (ECU) disposed in the vehicle; an external camera disposed on the vehicle and in communication with the ECU, the external camera arranged to detect a facial identity and a visual domain position of the user based on visual domain data of the user; and a wireless positioning transmitter disposed on the vehicle and in communication with the ECU, the wireless positioning transmitter arranged to detect an RF domain position of the vehicle based on RF domain data of the vehicle, the wireless positioning transmitter arranged to detect an RF domain position of the user based on RF domain data of the handheld device; and a cloud server remote from the vehicle and in communication with the ECU and the handheld device, the cloud server arranged to compare the visual domain position of the user with the RF domain position of the user by
a first equation,
|d.sub.visual.sup.t.sup.
and a second equation,
|φ.sub.visual.sup.t.sup.
10. The system of claim 9 wherein the ECU is arranged to monitor a location of the vehicle, wherein the cloud server is arranged to monitor a position of a handheld device of a user based on an intentional detection system, and wherein the cloud server is arranged to detect whether the handheld device is within the threshold distance from the vehicle.
11. The system of claim 10 wherein the cloud server is arranged to verify the facial identity of the user based on visual domain data of the user, wherein the cloud server is arranged to estimate the visual domain position of the user based on visual domain data of the user, wherein the cloud server is arranged to determine an RF domain position of the vehicle based on RF domain data of the vehicle, and wherein the cloud server is arranged to determine the RF domain position of the user based on RF domain data of the handheld device.
12. The system of claim 11 wherein the cloud server is arranged to detecting a non-user with a non-user device based on visual domain data of the non-user and RF domain data of the non-user device to define a visual domain position and an RF domain position of the non-user, wherein the cloud server is arranged to compare the visual domain position and the RF domain position of the user with the visual domain position and the RF domain position of the non-user by a third equation,
|d.sub.visual.sup.t.sup.
a fourth equation,
φ.sub.visual.sup.t.sup.
a fifth equation,
|d.sub.RF.sup.t.sup.
a sixth equation,
|φ.sub.RF.sup.t.sup.
13. The system of claim 9 wherein the onboard apparatus of the vehicle further comprises: an internal camera disposed in the vehicle and in communication with the ECU, the internal camera arranged to detect the facial identity and the visual domain position of the user based on visual domain data of the user when the user is in the vehicle, wherein the cloud server is arranged to verify the facial identity of the user based on visual domain data of the user when the user is in the vehicle.
14. The system of claim 9 wherein the cloud server is arrange to perform a sequence of actions within the vehicle based on personalized settings of the user after the vehicle is unlocked.
15. The system of claim 9 the onboard apparatus further comprises: an internal camera disposed in the vehicle and in communication with the ECU, the internal camera arranged to detect a companion-user having a companion-user device based on visual domain data of the companion-user and RF domain data of the companion-user device to define a visual domain position and an RF domain position of the companion-user; and wherein the cloud server is arranged to perform a sequence of actions within the vehicle based on personalized settings of the companion-user after unlocking the vehicle.
16. A method of unlocking a vehicle of a user having a handheld device based on visual domain and radio frequency (RF) domain technologies, the method comprising: monitoring a location of the vehicle and a position of the handheld device of a user based on an intentional detection system; detecting whether the handheld device is within a threshold distance from the vehicle; activating an on-board apparatus of the vehicle when the handheld device is within the threshold distance from the vehicle; verifying a facial identity of the user based on visual domain data of the user; estimating a visual domain position of the user based on visual domain data of the user; determining an RF domain position of the vehicle based on RF domain data of the vehicle; determining an RF domain position of the user based on RF domain data of the handheld device; comparing the visual domain position of the user with the RF domain position of the user by a first comparison,
|d.sub.visual.sup.t.sup.
and a second comparison,
|φ.sub.visual.sup.t.sup.
17. The method of claim 16 wherein the step of comparing further comprises: detecting a non-user having a non-user device based on visual domain data of the non-user and RF domain data of the non-user device to define a visual domain position and an RF domain position of the non-user; comparing the visual domain position and the RF domain position of the user with the visual domain position and the RF domain position of the non-user by a third equation,
|d.sub.visual.sup.t.sup.
a fourth equation,
φ.sub.visual.sup.t.sup.
a fifth equation,
|d.sub.RF.sup.t.sup.
a sixth equation,
|φ.sub.RF.sup.t.sup.
18. The method of claim 16 further including performing a sequence of actions within the vehicle based on personalized settings of the user after unlocking the vehicle.
19. The method of claim 16 further comprising: detecting a companion-user having a companion-user device based on visual domain data of the companion-user and RF domain data of the companion-user device to define a visual domain position and an RF domain position of the companion-user; and performing a sequence of actions within the vehicle based on personalized settings of the companion-user after unlocking the vehicle.
20. The method of claim 16 wherein verifying the facial identity of the user comprises: assessing image data quality of the facial identity of the user based on visual domain data of the user; determining a confidence threshold based on the image quality of the facial identity of the user; processing the visual domain data of the user with a facial recognition module to define a confidence score of the facial identity of the user; and confirming the facial identity of the user with an authentication module when the confidence score is greater than the confidence threshold.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) The drawings described herein are for illustration purposes only and are not intended to limit the scope of the present disclosure in any way.
(2)
(3)
(4)
(5)
(6)
DETAILED DESCRIPTION
(7) The following description is merely exemplary in nature and is not intended to limit the present disclosure, application, or uses.
(8) Aspects of the present disclosure provide a system and method of securely unlocking a vehicle of a user having a handheld device based on visual domain and wireless (e.g., radio frequency) domain technologies. The system and method more securely and more accurately validates a user of a vehicle when the user is within a threshold distance from the vehicle. The system and method compare visual domain data of the user and wireless domain data of the user's handheld device to more securely validate the user of the vehicle. Upon validation of the user, the system and method unlock the vehicle and activate user personalized settings in the vehicle.
(9)
(10) As shown, each of the external cameras 24, 25 and the wireless positioning transmitter 22 is in two-way communication with the ECU 20. That is, the external cameras 24, 25 may transmit and receive data signals 28, 29 respectively to and from the ECU 20. As described in greater detail below, the external camera 24 is arranged to detect a facial identity and a visual domain position of the user 14 based on visual domain data of the user 14. Moreover, the wireless positioning transmitter 22 may also transmit and receive data signals 30 respectively to and from the ECU 20 as described below. Further, the cloud server 18 may transmit and receive data signals 32 to and from the handheld device 15. Furthermore, the wireless positioning transmitter 22 may transmit and receive data signals 34 to and from the handheld device 15 as described in greater detail below. It is to be understood that any suitable cloud server, ECU, external camera, and wireless positioning camera may be used to achieve their respective functions as provided herein without departing from the spirit or scope of the present disclosure.
(11) As mentioned, the onboard apparatus 16 comprises the wireless positioning transmitter 22 disposed on the vehicle 12 and in communication with the ECU 20. The wireless positioning transmitter 22 is arranged to receive and transmit radio frequency (RF) domain position data of the vehicle 12 such that an RF domain position of the vehicle 12 may be detected by way of the wireless positioning transmitter 22, the ECU 20, or the cloud server 18 as desired. Moreover, the wireless positioning transmitter 22 is arranged to receive and transmit RF domain position data of the handheld device 15 such that an RF domain position of the user 14 may be detected via the wireless positioning transmitter 22, the ECU 20, or the cloud server 18 as desired.
(12) As shown, the cloud server 18 is disposed remotely from the vehicle 12. The cloud server is in communication with the ECU 20 and the handheld device 15 of the user 14. The handheld device 15 may be a smart phone, e.g., an iphone, or a keyfob.
(13) In one embodiment, when the user 14 is away from the vehicle 12, the system monitors a proximity of the user 14 relative to the vehicle 12. When the proximity of the user 14 relative to the vehicle 12 is within a predetermined distance, the on-board apparatus 16 is activated for use. For example, the ECU 20 is arranged to receive signals from the cloud server 18 to activate the on-board apparatus 16 when the handheld device 15 is within a threshold distance from the vehicle 12. To activate the on-board apparatus 16, the ECU 20 is arranged to monitor a location of the vehicle 12 based on an intentional detection system, e.g., GPS, and send GPS data signals of the vehicle 12 accordingly to the cloud server 18. Additionally, the cloud server 18 is arranged to monitor a position of the handheld device 15 of the user 14 based on an intentional detection system, e.g., GPS. Preferably, the cloud server 18 may receive GPS data signals from the handheld device 15 (e.g., via phone relative positioning and motion/intention detection modules) to monitor the position of the handheld device 15. With both the location of the vehicle 12 and the position of the handheld device 15 based on GPS, the cloud server 18 is arranged to detect whether the handheld device 15 is within the threshold distance from the vehicle 12. When the handheld device 12 is within the threshold distance from the vehicle 12, the cloud server sends a signal to the ECU 20 to activate the on-board apparatus 16 for operation.
(14) Referring to
(15) With the visual domain data and the RF domain data from the ECU 20, the cloud server 18 is arranged to verify the facial identity of the user 14 based on the visual domain data of the user 14. In one verification example, the cloud server 18 is arranged to assess image data quality of the facial identity of the user 14 based on the visual domain data of the user 14. The cloud server 18 is arranged to determine a confidence threshold based on an image quality of the facial identity of the user 14. If the image quality is relatively low, the confidence threshold may be lowered. Moreover, if the image quality is relatively high, the confidence threshold may be increased. By way of a facial recognition module, the cloud server 18 is arranged to process the visual domain data of the user 14 to define a confidence score of the facial identity of the user 14. By way of an authentication module, the cloud server 18 is arranged to confirm (verify) the facial identity of the user 14 when the confidence score is greater than the confidence threshold.
(16) Upon facial identity verification, the cloud server 18 is arranged to estimate the visual domain position of the user 14 based on visual domain data of the user 14 from the ECU 20. Additionally, the cloud server 18 is arranged to determine an RF domain position of the vehicle 12 based on the RF domain data of the vehicle 12. Furthermore, the cloud server 18 is arranged to determine the RF domain position of the user 14 based on the RF domain data of the handheld device 15.
(17) In this embodiment, the cloud server 18 is arranged to compare a first distance with a second distance. The first distance is a distance between the user 14 and the vehicle 12 based on visual domain data of the user 14. The second distance is a distance between the user 14 and the vehicle 12 based on RF domain data. Furthermore, a vehicle's local coordinate system may be used to describe the relation between the vehicle 12 and the user 14. It is to be understood that the vehicle local coordinate system may be a Cartesian coordinate system, a Spherical coordinate system, or any other suitable coordinate system without departing from the spirit or scope of the present disclosure.
(18) For example, the cloud server 18 is arranged to compare the visual domain position of the user 14 with the RF domain position of the user 14 by a first equation, |d.sub.visual.sup.t.sup.
(19) In this example, the term d.sub.visual.sup.t.sup.
(20) Moreover, the cloud server 18 is arranged to validate the user 14 when the first and second equations are true. Furthermore, the cloud server 18 is arranged to unlock the vehicle 12 when the user 14 is validated. For example, the cloud server 18 may send a command signal to the ECU 20 to unlock a door of the vehicle 12.
(21) For safety purposes, the cloud server 18 may be arranged to detect a non-user relative to the user 14 within the threshold distance and pause any steps of the system if the non-user is relatively close in proximity to the user. Such feature may be helpful to avoid potentially unsafe situations by a malicious non-user. Thus, the cloud server may be arranged to check the distance between the user and non-user. Moreover, if the non-user is detected to be within a minimum distance from the user, the cloud server may be arranged to pause any steps of unlocking the vehicle. The cloud server may also be arranged to contact authorities if desired.
(22) For example, the cloud server 18 may be arranged to detect the non-user 42 (
(23) Additionally, the wireless positioning transmitter 22 may sense the handheld device of the non-user 42 and may be arranged to receive RF data from the non-user's handheld device (via positioning module or motion/intention detection module) for the RF domain position of the non-user. In accordance with the RF data from the non-user's handheld device, a non-user RF domain data signal is sent to the ECU 20 which in turn transmits non-user RF domain data to the cloud server 18.
(24) In this example, the cloud server 18 is arranged to compare the visual domain position and the RF domain position of the user 14 with the visual domain position and the RF domain position of the non-user 42 by a third equation |d.sub.visual.sup.t.sup.
(25) In this embodiment, d.sub.Visual.sup.t.sup.
(26) Furthermore, the cloud server 18 is arranged to proceed to validating the user 14 when the third, fourth, fifth, and sixth equations are true. However, if the any of the third, fourth, fifth, or sixth equations is not true, then it may be determined that the non-user 42 may be within the minimum distance relative to the user 14. Hence, any steps of unlocking the vehicle 12 may be unsafe or not secure. Therefore, all steps of unlocking the vehicle may be temporarily paused.
(27) In another embodiment, the user's facial identity may be repeatedly verified when the user 14 is in the vehicle 12. Such verification may be helpful for added security and authorization especially in situations where a user may frequently depart a vehicle temporarily. For example, when a vehicle is used in transportation and courier industries, facial identity verification may be desired for added security and authorization of the vehicle.
(28) As an example and shown in
(29) Upon unlocking the vehicle 12, the cloud server 18 is arranged to perform a sequence of actions within the vehicle 12 based on personalized settings of the user 14 after the vehicle 12 is unlocked. For example, the user may have preset personalized settings saved in the ECU 20 and such personalized settings may be activated within the vehicle 12. Such personalized settings may include but is not limited to temperature control, audio, stereo, seat, steering wheel, and driving mode settings.
(30) For convenience and added security purposes, the system may be arranged to detect and verify a companion-user when in the vehicle. Upon detection and verification, the system 10 may be arranged to perform a sequence of actions with the vehicle 12 based on personalized settings of the companion-user. For example, the companion-user may have preset personalized settings saved in the ECU 20 and such personalized settings may be activated within the vehicle 12. Such personalized settings may include but is not limited to temperature control, audio, stereo, seat, steering wheel, and driving mode settings.
(31) As an example thereof, the internal camera may be arranged to detect a companion-user having a companion-user device based on visual domain data and RF domain data of the companion-user device to define a visual domain position and an RF domain position of the companion-user. In this example, the internal camera is arranged to detect the facial identity and the visual domain position of the companion-user based on visual domain data of the companion-user in the vehicle 12. Additionally, the internal camera may sense the companion-user and capture an image of the companion-user for the visual domain position thereof when in the vehicle 12. In accordance with the image of the companion-user, a companion-user visual domain data signal is sent to the ECU 20 which in turn transmits companion-user visual domain data to the cloud server 18. Preferably, the cloud server 18 is arranged to verify the facial identity of the companion-user based on visual domain data of the companion-user in the vehicle 12.
(32) Upon companion-user verification, the cloud server 18 is arranged to perform a sequence of actions within the vehicle 12 based on personalized settings of the companion-user. As mentioned above, the companion-user may have preset personalized settings saved in the ECU 20 and such personalized settings may be activated within the vehicle 12. Such personalized settings may include but is not limited to temperature control, audio, stereo, seat, steering wheel, and driving mode settings.
(33) It is to be understood that at least one of the cloud server 18 and the on-board apparatus 16, e.g., the ECU 20, may comprise a number of algorithms and modules having algorithms to achieve the tasks provided herein. For example, the cloud server and the on-board apparatus may comprise a user/vehicle registration module to validate users and their vehicles. Moreover, the cloud server and on-board apparatus may comprise a resource discovery module to monitor and record the location of the user's vehicle and handheld device 15. Further, the cloud server and on-board apparatus may comprise an image processing (including face recognition and positioning) module to recognize and position the user based on visual domain data. Moreover, the cloud server and on-board apparatus may comprise relative positioning sensor fusion module to compare visual domain positions with RF domain positions of any of the user, non-user, and companion-user. Additionally, the cloud server and the on-board apparatus may comprise a user identification and personalization module to store user identification and personalized settings thereof.
(34) Moreover, the cloud server and the on-board apparatus may comprise a WiFi/UWB positioning module to detect and send wireless domain data of any of the vehicle, user device, non-user device, and companion-user device. In addition, the cloud server and the on-board apparatus may comprise a quality assessment module to assess the image quality of any of the user, non-user, and companion-user. Further, the cloud server and the on-board apparatus may comprise an authentication module to detect and verify faces of any of the user, non-user, and companion-user based on visual domain data. Furthermore, the cloud server and the on-board apparatus may comprise a face recognition and face positioning module to recognize and position of any of the user, non-user, and companion-user based on visual domain data. Other modules and algorithms may be used without departing from the scope or spirit of the present disclosure.
(35) It is to be understood that the handheld device of the user, non-user, and companion-user may comprise a number of algorithms and modules having algorithms to achieve the tasks provided herein. For example, the handheld device may comprise a vehicle/phone relative positioning module for positioning the vehicle and the handheld device. Moreover, the handheld device may comprise a motion/intention detection module for detecting and sensing motion of the vehicle or device.
(36)
(37) In this aspect, d.sub.visual.sup.t.sup.
(38) Moreover, the method 110 further comprises a step 116 of validating the user 14 when the first and second equations are true and a step 118 of unlocking the vehicle 12 after the user 14 is validated.
(39)
(40) In this aspect, the method 210 further comprises a step 220 of estimating a visual domain position of the user 14 based on visual domain data of the user 14, a step 222 of determining an RF domain position of the vehicle 12 based on RF domain data of the vehicle 12, and a step 224 of determining an RF domain position of the user 14 based on RF domain data of the handheld device 15.
(41) Moreover, in this aspect, the method 210 further comprises a step 226 of comparing the visual domain position of the user 14 with the RF domain position of the user 14 by a first equation, |d.sub.visual.sup.t.sup.
(42) Furthermore, the method 210 further comprises a step 228 of validating the user 14 when the first and second comparisons are true and a step 230 of unlocking the vehicle 12 after the user 14 is validated.
(43) The description of the present disclosure is merely exemplary in nature and variations that do not depart from the gist of the present disclosure are intended to be within the scope of the present disclosure. Such variations are not to be regarded as a departure from the spirit and scope of the present disclosure.