Gesture recognition method and gesture recognition system
10817712 · 2020-10-27
Assignee
Inventors
- Tsung-Ming Tai (New Taipei, TW)
- Yun-Jie Jhang (Taoyuan, TW)
- Wen-Jyi Hwang (Taipei, TW)
- Chun-Hsuan KUO (San Diego, CA, US)
Cpc classification
G06F18/217
PHYSICS
G06F17/18
PHYSICS
H03B21/02
ELECTRICITY
G06F18/2155
PHYSICS
G01S13/50
PHYSICS
G06F18/214
PHYSICS
G06F9/5027
PHYSICS
G06F3/017
PHYSICS
G01S7/415
PHYSICS
International classification
G06F9/50
PHYSICS
G01S13/50
PHYSICS
G01S13/58
PHYSICS
G01S7/41
PHYSICS
H03B21/02
ELECTRICITY
G06F17/18
PHYSICS
Abstract
A gesture recognition system executes a gesture recognition method. The gesture recognition method includes steps of: receiving a training signal; selecting one of the sensing frames of the sensing signal; generating a sensing map; selecting a cell having the max-amplitude; determining a frame amplitude, a frame phase, and a frame range of the selected one of the sensing frames; setting the frame amplitudes, the frame phases, and the frame ranges of all of the sensing frames to input data of a neural network to classify a gesture event. The present invention just uses a few data to be the input data of the neural network. Therefore, the neural network may not require high computational complexity, the gesture recognition system may decrease the calculation load of the processing unit, and the gesture recognition function may not influence a normal operation of a smart device.
Claims
1. A gesture recognition method, comprising a performing procedure; wherein the performing procedure is executed by a central processing unit of a smart device, and comprises: receiving a sensing signal; wherein the sensing signal comprises a plurality of sensing frames; wherein the sensing signal is a Range Doppler Image (RDI) signal generated by a Doppler radar; selecting one of the sensing frames of the sensing signal; generating a sensing map according to the selected one of the sensing frames; wherein the sensing map comprises a plurality of chirps, each of the chirps comprises a plurality of cells, and each of the cells has an amplitude and a phase; selecting the cell having the max-amplitude in each of the chirps as an interested cell; determining a frame amplitude, a frame phase, and a frame range of the selected one of the sensing frames according to the amplitudes and the phases of the interested cells of the chirps; determining whether the frame amplitudes, the frame phases, and the frame ranges of all of the sensing frames of the sensing signal are determined; when the frame amplitudes, the frame phases, and the frame ranges of all of the sensing frames of the sensing signal are determined, setting the frame amplitudes, the frame phases, and the frame ranges of all of the sensing frames of the sensing signal to input data of a neural network to classify a gesture event.
2. The gesture recognition method as claimed in claim 1, wherein the frame amplitude, the frame phase, and the frame range the selected one of the sensing frames are determined by: calculating an average amplitude by averaging the amplitudes of the interested cells of the chirps; setting the average amplitude to the frame amplitude of the selected one of the sensing frames; calculating an average phase by averaging the phases of the interested cells of the chirps; setting the average phase to the frame phase of the selected one of the sensing frames; calculating ranges of the interested cells according to the amplitudes of the interested cells; calculating an average range by averaging ranges of the interested cells of the chirps; and setting the average range to the frame range of the selected one of the sensing frames.
3. The gesture recognition method as claimed in claim 1, wherein when the amplitude, the phase, and the range of all of the sensing frames of the sensing signal are not determined, selecting another one of the sensing frames of the sensing signal, and generating the sensing map according to the selected one of the sensing frames again.
4. The gesture recognition method as claimed in claim 1, wherein the sensing map is generated by transforming the selected one of the sensing frames using a Fast Fourier Transform.
5. The gesture recognition method as claimed in claim 1, wherein the neural network is a recurrent neural network (RNN).
6. The gesture recognition method as claimed in claim 5, wherein the recurrent neural network is a Long Short Memory (LSTM) network, a Gated Recurrent Unit (GRU), a Simple Gated Unit (SGU), or a Minimal Gated Unit (MGU).
7. A gesture recognition system, comprising a performing module; wherein the performing module comprises: a sensing unit, receiving a sensing signal; wherein the sensing signal comprises a plurality of sensing frames; wherein the sensing unit is a Doppler radar, and the sensing signal is a Range Doppler Image (RDI) signal generated by the Doppler radar; a memory unit, storing a neural network; a processing unit, electrically connected to the sensing unit and the memory unit, and receiving the sensing signal from the sensing unit; wherein the processing unit selects one of the sensing frames of the sensing signal, and generates a sensing map according to the selected one of the sensing frames; wherein the sensing map comprises a plurality of chirps, each of the chirps comprises a plurality of cells, and each of the cells has an amplitude and a phase; wherein the processing unit is a central processing unit of a smart device; wherein the processing unit selects the cell having the max-amplitude in each of the chirps as an interested cell, and determines a frame amplitude, a frame phase, and a frame range of the selected one of the sensing frames according to the amplitudes and the phases of the interested cells of the chirps; wherein the processing unit further determines whether the frame amplitudes, the frame phases, and the frame ranges of all of the sensing frames of the sensing signal are determined; wherein when the frame amplitudes, the frame phases, and the frame ranges of all of the sensing frames of the sensing signal are determined, the processing unit loads the neural network from the memory, and sets the frame amplitudes, the frame phases, and the frame ranges of all of the sensing frames of the sensing signal to input data of the neural network to classify a gesture event.
8. The gesture recognition system as claimed in claim 7, wherein the processing unit determines the frame amplitude, the frame phase, and the frame range of the selected one of the sensing frames by steps of: calculating an average amplitude by averaging the amplitudes of the interested cells of the chirps; setting the average amplitude to the frame amplitude of the selected one of the sensing frames; calculating an average phase by averaging the phases of the interested cells of the chirps; setting the average phase to the frame phase of the selected one of the sensing frames; calculating ranges of the interested cells according to the amplitudes of the interested cells; calculating an average range by averaging ranges of the interested cells of the chirps; setting the average range to the frame range of the selected one of the sensing frames.
9. The gesture recognition system as claimed in claim 7, wherein when the amplitude, the phase, and the range of all of the sensing frames of the sensing signal are not determined, the processing unit selects another one of the sensing frames of the sensing signal, and generates the sensing map according to the selected one of the sensing frames again.
10. The gesture recognition system as claimed in claim 7, wherein the processing unit transforms the selected one of the sensing frames by a Fast Fourier Transform to generate the sensing map.
11. The gesture recognition system as claimed in claim 7, wherein the neural network is a recurrent neural network (RNN).
12. The gesture recognition system as claimed in claim 11, wherein the recurrent neural network is a Long Short Memory (LSTM) network, a Gated Recurrent Unit (GRU), a Simple Gated Unit (SGU), or a Minimal Gated Unit (MGU).
Description
BRIEF DESCRIPTION OF THE DRAWINGS
(1)
(2)
(3)
(4)
DETAILED DESCRIPTION OF THE INVENTION
(5) With reference to
(6) receiving a sensing signal; wherein the sensing signal includes a plurality of sensing frames (S101);
(7) selecting one of the sensing frames of the sensing signal (S102);
(8) generating a sensing map according to the selected one of the sensing frames (S103); wherein the sensing map includes a plurality of chirps, each of the chirps includes a plurality of cells, and each of the cells has an amplitude and a phase;
(9) selecting the cell having the max-amplitude in each of the chirps as an interested cell (S104);
(10) determining a frame amplitude, a frame phase, and a frame range of the selected one of the sensing frames according to the amplitudes and the phases of the interested cells of the chirps (S105);
(11) determining whether the frame amplitudes, the frame phases, and the frame ranges of all of the sensing frames of the sensing signal are determined (S106);
(12) when the frame amplitudes, the frame phases, and the frame ranges of all of the sensing frames of the sensing signal are determined, setting the frame amplitudes, the frame phases, and the frame ranges of all of the sensing frames of the sensing signal to input data of a neural network to classify a gesture event (S107).
(13) Moreover, the gesture recognition method further includes a step of:
(14) when the amplitude, the phase, and the range of all of the sensing frames of the sensing signal are not determined, selecting another one of the sensing frames of the sensing signal (S108), and generating the sensing map according to the selected one of the sensing frames again (S103).
(15) With reference to
(16) The sensing unit 11 receives a sensing signal, and the sensing signal includes a plurality of sensing frames.
(17) The memory unit 13 stores a neural network.
(18) The processing unit 12 is electrically connected to the sensing unit 11 and the memory unit 13, and receives the sensing signal from the sensing unit 11. The processing unit 12 selects one of the sensing frames of the sensing signal, and generates a sensing map according to the selected one of the sensing frames; wherein the sensing map includes a plurality of chirps, each of the chirps includes a plurality of cells, and each of the cells has an amplitude and a phase.
(19) The processing unit 12 selects the cell having the max-amplitude in each of the chirps as an interested cell, and determines a frame amplitude, a frame phase, and a frame range of the selected one of the sensing frames according to the amplitudes and the phases of the interested cells of the chirps.
(20) The processing unit 12 further determines whether the frame amplitudes, the frame phases, and the frame ranges of all of the sensing frames of the sensing signal are determined.
(21) When the frame amplitudes, the frame phases, and the frame ranges of all of the sensing frames of the sensing signal are determined, the processing unit 12 loads the neural network from the memory 13, and sets the frame amplitudes, the frame phases, and the frame ranges of all of the sensing frames of the sensing signal to input data of the neural network to classify a gesture event.
(22) Further, when the amplitude, the phase, and the range of all of the sensing frames of the sensing signal are not determined, the processing unit 12 selects another one of the sensing frames of the sensing signal, and generates the sensing map according to the selected one of the sensing frames again.
(23) Since each of the frames just has one frame amplitude, one frame phase, and one frame range, the neural network can classify the gesture event according to a few input data. For example, when the amount of the frames is 10, the present invention may use 30 parameters to be the input data of the neural network to classify the gesture event.
(24) The present invention may not use all of parameters of the cells of the frames to be the input data of the neural network. The neural network may not require high computational complexity, and the gesture recognition system of the present invention may decrease a calculation load of the processing unit 12. Therefore, when the present invention uses a central processing unit of a smart device to be the processing unit 12 of the gesture recognition system for executing the performing procedure, the gesture recognition function may not influence performance of the smart device.
(25) Namely, the gesture recognition system may be easily integrated into the smart device, such as a smart phone, a tablet, or a computer.
(26) Further, with reference to
(27) calculating an average amplitude by averaging the amplitudes of the interested cells of the chirps (S1051);
(28) setting the average amplitude to the frame amplitude of the selected one of the sensing frames (S1052);
(29) calculating an average phase by averaging the phases of the interested cells of the chirps (S1053);
(30) setting the average phase to the frame phase of the selected one of the sensing frames (S1054);
(31) calculating ranges of the interested cells according to the amplitudes of the interested cells (S1055);
(32) calculating an average range by averaging ranges of the interested cells of the chirps (S1056);
(33) setting the average range to the frame range of the selected one of the sensing frames (S1057).
(34) In the embodiment, the sensing map is generated by transforming the selected one of the sensing frames using the Fast Fourier Transform (FFT), the neural network is a recurrent neural network (RNN), and the recurrent neural network is a Long Short Memory (LSTM) network, a Gated Recurrent Unit (GRU), a Simple Gated Unit (SGU), or a Minimal Gated Unit (MGU). Further, the sensing signal is a Range Doppler Image (RDI) signal.
(35) With reference to
(36) receiving a training signal; wherein the training signal comprises a plurality of training frames (S401);
(37) selecting one of the training frames of the training signal (S402);
(38) generating a training map according to the selected one of the training frames (S403); wherein the training map comprises a plurality of chirps, each of the chirps comprises a plurality of cells, and each of the cells has an amplitude and a phase;
(39) selecting the cell having the max-amplitude in each of the chirps as an interested cell (S404);
(40) determining a frame amplitude, a frame phase, and a frame range of the selected one of the training frames according to the amplitudes and the phases of the interested cells of the chirps (S405);
(41) determining whether the frame amplitudes, the frame phases, and the frame ranges of all of the training frames of the training signal are determined (S406);
(42) when the frame amplitudes, the frame phases, and the frame ranges of all of the training frames of the training signal are determined, setting the frame amplitudes, the frame phases, and the frame ranges of all of the training frames of the training signal to training data of the neural network to train the neural network (S407);
(43) when the frame amplitudes, the frame phases, and the frame ranges of all of the training frames of the training signal are not determined, selecting another one of the training frames of the training signal (S408), and generating the training map according to the selected one of the training frames again (S403).
(44) Further with reference to
(45) For example, the training module 20 stores a training neural network, and is electrically connected to the performing module 10 to receive a training signal from the sensing unit 11 of the performing module 10. The training signal comprises a plurality of training frames.
(46) The training module 20 selects one of the training frames of the training signal, and generates a training map according to the selected one of the training frames. The training map comprises a plurality of chirps, each of the chirps comprises a plurality of cells, and each of the cells has an amplitude and a phase.
(47) The training module 20 selects the cell having the max-amplitude in each of the chirps as an interested cell, and determines a frame amplitude, a frame phase, and a frame range of the selected one of the training frames according to the amplitudes and the phases of the interested cells of the chirps.
(48) The training module 20 further determines whether the frame amplitudes, the frame phases, and the frame ranges of all of the training frames of the training signal are determined;
(49) When the frame amplitudes, the frame phases, and the frame ranges of all of the training frames of the training signal are determined, the training module 20 sets the frame amplitudes, the frame phases, and the frame ranges of all of the training frames of the training signal to training data of the training neural network to train the training neural network, and the training module 20 further updates the neural network of the performing module 10 by the training neural network.
(50) When the frame amplitudes, the frame phases, and the frame ranges of all of the training frames of the training signal are not determined, the training module 20 selects another one of the training frames of the training signal, and generates the training map according to the selected one of the training frames again.
(51) In the above embodiment, the training map is generated by transforming the selected one of the training frames using the Fast Fourier Transform (FFT), the training neural network of the training module 20 is a recurrent neural network (RNN), and the recurrent neural network is a Long Short Memory (LSTM) network, a Gated Recurrent Unit (GRU), a Simple Gated Unit (SGU), or a Minimal Gated Unit (MGU). Further, the training signal is a Range Doppler Image (RDI) signal.
(52) Even though numerous characteristics and advantages of the present invention have been set forth in the foregoing description, together with details of the structure and function of the invention, the disclosure is illustrative only. Changes may be made in detail, especially in matters of shape, size, and arrangement of parts within the principles of the invention to the full extent indicated by the broad general meaning of the terms in which the appended claims are expressed.