Eye-tracking communication methods and systems
11612342 · 2023-03-28
Assignee
Inventors
Cpc classification
G06F3/015
PHYSICS
G06F3/017
PHYSICS
A61B5/0205
HUMAN NECESSITIES
G06F3/0236
PHYSICS
International classification
A61B5/16
HUMAN NECESSITIES
A61B5/00
HUMAN NECESSITIES
A61B5/0205
HUMAN NECESSITIES
Abstract
Provided is a control system that interfaces with an individual through tracking the eyes and/or tracking other physiological signals generated by an individual. The system, is configured to classify the captured eye images into gestures, that emulate a joystick-like control of the computer. These gestures permit the user to operate, for instance a computer or a system with menu items.
Claims
1. An eye tracking-based system, comprising: a camera operable for continuously capturing images of one or both of the user's eye and eyelid and generating image data representative thereof; a first output module, a computerized or process-driven module; and a control unit in data communication with the camera and with the first output module; wherein the control unit is configured for receiving and processing said image data to identify at least one of pupil position and eyelid movement, and to classify the eye image into gestures based on pupil position and pupil presence duration within an area of a threshold map, said processing further comprises determining whenever the pupil area touches a border or is tangent to a border of the threshold map, to thereby define the gesture type, selected from one or more of pupil position, sequence of pupil positions, and sequences of eyelid blinks and generating gesture data, operating a hierarchical user-selectable menu items to permit the user to navigate through and select menu items by said gestures data, and for driving the first output module to present the menu items to the user; and wherein the first output module is configured for providing the user with audio presentation of a time-based prompt menu for selection of items with a predefined gesture, wherein a first menu item is announced in time t.sub.1 and a second menu item is announced in t.sub.2, and a third menu item is announced in t.sub.3, the first menu item is selected if said predefined gesture is made in a time between t.sub.1 and t.sub.2 and the second menu item is selected if said predefined gesture is made in a time between t.sub.2 and t.sub.3.
2. The system of claim 1, wherein said camera is carried on a holder attachable to the user's head.
3. The system of claim 1, comprising a driver for a second output module for outputting data representative of selected items.
4. The system of claim 3, wherein the second output module is configured for outputting an alert.
5. The system of claim 4, wherein at least one gesture triggers said alert.
6. The system of claim 3, wherein the second output module is configured for outputting an alert to a care-giver.
7. An eye tracking-based system, comprising: a camera operable for continuously capturing images of one or both of the user's eye and eyelid and generating image data representative thereof; a first output module, a computerized or process-driven module; and a control unit in data communication with the camera and with the first output module; wherein the control unit is configured for receiving and processing said image data to identify at least one of pupil position and eyelid movement, and to classify these into gestures comprising one or more of pupil position, sequence of pupil positions, and sequences of eyelid blinks and generating gesture data, operating a hierarchical user-selectable menu items to permit the user to navigate through and select menu items by said gesture data, and for driving the first output module to present the menu items to the user; and wherein the first output module is configured for providing the user with audio presentation of a time-based prompt menu for selection of items with a predefined gesture, wherein a first menu item is announced in time t.sub.1 and a second menu item is announced in t.sub.2, and a third menu item is announced in t.sub.3, the first menu item is selected if said predefined gesture is made in a time between t.sub.1 and t.sub.2 and the second menu item is selected if said predefined gesture is made in a time between t.sub.2 and t.sub.3.
8. An eye tracking-based system, comprising: a camera operable for continuously capturing images of one or both of the user's eye and eyelid and generating image data representative thereof; a first output module, a computerized or process-driven module; and a control unit in data communication with the camera and with the first output module; wherein the control unit is configured for receiving and processing said image data to identify at least one of pupil position and eyelid movement, and to classify the eye image into gestures based on pupil position and pupil presence duration within an area of a threshold map, said processing further comprises determining whenever the pupil area touches a border or is tangent to a border of the threshold map, to thereby define the gesture type, selected from one or more of pupil position, sequence of pupil positions, and sequences of eyelid blinks and generating gesture data, operating a hierarchical user-selectable menu items to permit the user to navigate through and select menu items by said gestures data, and for driving the first output module to present the menu items to the user; and wherein the first output module is configured for providing the user with audio presentation of a time-based prompt menu for selection of items with a predefined gesture, wherein a first menu item is announced in time t.sub.1 and a second menu item is announced in t.sub.2, and a third menu item is announced in t.sub.3, wherein if said predefined gesture is made in a time t, the first menu item is selected if t is between t.sub.1 and t.sub.2 and the second menu item is selected if t is between t.sub.2 and t.sub.3.
9. An eye tracking-based system, comprising: a camera operable for continuously capturing images of one or both of the user's eye and eyelid and generating image data representative thereof; a first output module, a computerized or process-driven module; and a control unit in data communication with the camera and with the first output module; wherein the control unit is configured for receiving and processing said image data to identify at least one of pupil position and eyelid movement, and to classify the eye image into gestures based on pupil position and pupil presence duration within an area of a threshold map, said processing further comprises determining whenever the pupil area touches a border or is tangent to a border of the threshold map, to thereby define the gesture type, selected from one or more of pupil position, sequence of pupil positions, and sequences of eyelid blinks and generating gesture data, operating a hierarchical user-selectable menu items to permit the user to navigate through and select menu items by said gestures data, and for driving the first output module to present the menu items to the user; and wherein the first output module is configured for providing the user with audio presentation of a time-based prompt menu for selection of items with a predefined gesture, wherein a first menu item is announced in time t.sub.1 and a second menu item is announced in t.sub.2, and a third menu item is announced in t.sub.3, wherein t.sub.2 is a time after t.sub.1 and t.sub.3 is a time after t.sub.2, the first menu item is selected if said predefined gesture is made in a time frame between t.sub.1 and t.sub.2 and the second menu item is selected if said predefined gesture is made in a time frame between t.sub.2 and t.sub.3.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1) In order to better understand the subject matter that is disclosed herein and to exemplify how it may be carried out in practice, embodiments will now be described, by way of non-limiting example only, with reference to the accompanying drawings, in which:
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
(10)
(11)
(12)
(13)
(14)
DETAILED DESCRIPTION OF EMBODIMENTS
(15) Reference is first made to
(16) Control unit 104 includes also a processor 110 that is configured for receiving and processing image data from the camera 102 and for identifying at least one of pupil position and eyelid movement and to classify these into gestures comprising one or more of pupil position, sequence of pupil positions, and sequences of eyelid blinks and generating gesture data. The processor 110 is also configured for driving the menu generator 112 which, through the action of actuator module 108 drives the presentation of the menu to the user. This permits the user to navigate through and select menu items by said gesture data.
(17)
(18) In
(19) The system of
(20) Reference is now being made to
(21) Reference is now being made to
(22) At any time and in any layer of the menu, when the user will make a predefined gestures sequence PGS, it will trigger a predefined action such as outputting an emergency alert for a caregiver e.g. by voice alert through a speaker, textual alert to a mobile device, alerting a medical center or any combination thereof. The predefined gestures sequence PGS may be configured according the user's will, for example it can be a sequence of 3 or 4 blinks, a sequence of up gesture UG, down gesture DG, up gesture UG and down gesture DG, or any other desired sequence.
(23)
(24) In order to improve classification of the gestures, the system may be trained by a machine/deep learning algorithm. First, the system is received with labeled gestures images (Blink, Center, Up, Down, Right, Left) to gather initial dataset. Then, the system go through a training session with a set of training images. During this training session the system, namely the neural network of the system, learns how to recognize each of the categories in the labeled images. When the present model makes a mistake, it corrects itself and improves. When the training session of the network is over, a testing set of images is received and processed by the system to check the new model of classification. The classification made by the system is compared with the ground-truth labels of the testing set and the number of correct classifications can be computed and values of precision, recall, and f-measure, which are used to quantify the performance of such a network can be obtained.
(25) A schematic illustration of an assistive communication eye tracking-based system is provided by
(26) In clinical trials carried out by the inventors of the present application it demonstrated that patients who were able to comfortably control the system following a brief, several minutes trial. As a non-limiting example provide below as Table 1, in a clinical trial held at Rambem Hospital, Israel, studying the “call for help” function required an average training time of 1.12 minutes, studying to communicate predetermined set of sentences required an average training time of 6.44 minutes, and free-text letter by letter communication using a mobile screen required an average training time of 11.08 minutes.
(27) TABLE-US-00001 TABLE 1 Communication Average training type time (minutes) “Call for help” 1.12 A sentence 6.44 Free text 11.08
(28) A non-limiting embodiment of the joystick-like gesture classification is illustrated in
(29)