DRONE FLIGHT CONTROL CENTER AND PILOT MONITOR
20250164998 ยท 2025-05-22
Inventors
Cpc classification
G05D1/223
PHYSICS
International classification
Abstract
An sUAS operations device that allows a pilot to fly an sUAS. The device includes a sensor interface to receive signals from a sensor attached to the pilot to monitor physiological responses of the pilot during the flight. The system records a video of the flight and synchronizes it with the sensor feedback to allow correlation of the stress levels with the flight. The system also may provide feedback to the pilot during flight.
Claims
1. A ground control station for an unmanned aerial vehicle (UAV), comprising: controls to allow a user to control flight of a UAV; a display to allow the user to view at least one of a video feed of the UAV and a video feed from the drone; one or more ports to allow the station to receive signals from one or more sensors connected to the user and one or more video feeds; a memory to store the one or more video feeds and signals received through the one or more ports; and one or more processors configured to execute code that causes the one more processors to: apply artificial intelligence to the signals and the one or more video feeds to provide an incident prediction; and notify the user if the incident prediction is above a threshold indicating a likelihood of an occurrence of an incident.
2. The ground control station as claimed in claim 1, further comprising a memory.
3. The ground control station as claimed in claim 2, wherein the one or more processors are further configured to execute code to store at least one of the signals from the one or more sensors and the one or more video feeds.
4. The ground control station as claimed in claim 1, wherein the one or more sensors comprise a wrist-mounted sensor, a sensor positioned on the ground control station to contact at least one hand of the user.
5. The ground control station as claimed in claim 1, wherein the one or more video feeds comprise one or more of a video feed from the UAV, a video feed of a view of the user, and a video feed of the user.
6. The ground control station as claimed in claim 1, wherein the one or more processors are further configured to execute code that causes the one or more processors to train the artificial intelligence.
7. The ground control station as claimed in claim 6, wherein the code that causes the one or more processors to train the artificial intelligence comprises code that causes the one or more processors to train the artificial intelligence on one or more dataset comprised of signals, video feeds, and results.
8. A ground control system, comprising: one or mores sensors configured to be in contact with a pilot; controls to allow a user to control flight of a UAV; one or more cameras, at least one camera attached to the UAV; a display to allow the user to view at least one of a graphic of the UAV and video from the drone; one or more ports to allow the station to receive signals from the one or more sensors connected to the user and one or more video feeds; a memory to store the one or more video feeds and signals received through the one or more ports; and one or more processors configured to execute code that causes the one more processors to: apply artificial intelligence to the signals and the one or more video feeds to provide an incident prediction; and notify the user if the incident prediction is above a threshold indicating a likelihood of an occurrence of an incident.
9. The ground control system as claimed in claim 8, wherein the one or more sensors comprise one or more of heart rate monitors, galvanic skin sensors, and EEGs.
10. The ground control system as claimed in claim 8, wherein the one or more sensors are configured to reside on one or more of a wrist, ear, and hand of the pilot.
11. The ground control system as claimed in claim 8, wherein the one or more cameras include additional cameras on one or more of the ground control station and near eyes of the pilot.
12. The ground control station as claimed in claim 8, further comprising a memory.
13. The ground control station as claimed in claim 12, wherein the one or more processors are further configured to execute code to store at least one of the signals from the one or more sensors and the one or more video feeds.
14. The ground control system as claimed in claim 8, wherein the one or more processors are further configured to execute code that causes the one or more processors train the artificial intelligence.
15. The ground control system as claimed in claim 14, wherein the code that causes the one or more processors to train the artificial intelligence comprises code that causes the one or more processors to train the artificial intelligence on one or more dataset comprised of signals, video feeds, and results.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0007]
[0008]
[0009]
DETAILED DESCRIPTION OF THE EMBODIMENTS
[0010] The embodiments here attached sensor(s) to a drone or other unmanned aerial vehicle (UAV) pilot. As used here, the term UAV includes UAVs, small UAVs (sUAV) and small Uncrewed Aircraft Systems, (sUAS) and any other kind of piloted drone. One embodiment involves a UAV pilot monitoring system, and the system includes the UAV, a ground control station (GCS), one or more sensors, and one or more video cameras.
[0011] Examples of sensors include a heart rate monitor, including heart rate variability monitor (HRV), galvanic skin response sensors, electroencephalogram (EEG) electrodes, etc. The resulting sensor tracking provides the pilot and other systems with feedback as to pilot's stress. The sensors are in physical contact with the pilot. The physical contact may result from the sensors being attached to the pilot, or the pilot touching the sensors, such as when the pilot handles the GCS, as examples without limitation. The sensors send signals related to the pilot's physical condition, such as their heart rhythm and heart rhythm coherence, the pilot's galvanic skin response, EEGs, etc.
[0012] The video cameras produce one or more video feeds to be displayed by one or more displays on the GCS. The one or more video feeds may include a video feed from the UAV showing a view that the UAV sees, a video feed of the pilot's view, obtained from a video camera located near the pilot's eyes, and a video feed of the pilot, more than likely obtained from a camera located at or near the GCS.
[0013]
[0014] The pilot monitoring system embodiment of
[0015]
[0016]
[0017] The GCS of these embodiments shows a self-contained GCS. One should note that this may comprise the most useful embodiment, as it does not rely upon a separate computing device. The GCS receives the signals from the sensor(s) and the one or more video feeds and applies an artificial intelligence (AI) machine learning model to the inputs. The GCS may synchronize the different video feeds and the signals using time stamps of their reception. The AI model will analyze the signals and video feeds to determine the likelihood of an incident related to the pilot performance and stress. The AI model may take many forms, including convolutional neural network (CNN), long short-term memory (LSTM) networks, radial basis functions (RBF) neural network, artificial neural network (ANN), recurrent neural network (RNN), as examples.
[0018] Prior to deployment in a training environment, the AI model will undergo training. The training may rely upon stored data sets comprised of the signals and video feeds associated with pilot-related incidents. These data sets may be collected over time during training simulations or actual training operations. The feeds before the incident and the resulting incident can train the model to recognize characteristics of the feeds that preceded the accidents. As the model undergoes training, it develops a capacity to view new video and sensor feeds and make predictions as to the possibility of a pilot incident.
[0019] When deployed, the system will take the signals and the video feeds and makes predictions of a possibility of an imminent incident. As the GCS evaluates the feeds and signals, the prediction may change. The GCS would then provide an indicator to the pilot. For example, the display may have a region in which a bar or a series of lights could provide instant feedback to the pilot. For example, as the AI model evaluates the inputs and finds no risk, it may display a green light on the display or as a separate LED or other light. If the probability of an incident increases, the light could turn yellow to give the pilot a chance to adjust and adapt their behavior to correct whatever issues are causing the risk. When the light turns red, the system could automatically shut down, or would send the pilot a message that the pilot needs to stop.
[0020] The GCS has an architecture somewhat similar to a computing device, except that it has connections or ports that allow the GCS to communicate with the UAV and the sensors and/or cameras that reside on the user. As mentioned above, the GCS will receive signals from the sensor(s) and one or more video feeds. The video feeds will include the video from the drone, and possibly a video feed of the user from a camera mounted on or near the GCS, and possibly a video feed from a camera mounted near the pilot's eyes.
[0021]
[0022] In
[0023] Utilizing this unit and the training will reduce incidents and accidents related to human factors. Increasing training for the remote pilot to handle emergencies, either in simulations or actual flights, will reduce these accidents. Presently, this system and methodology remains the only known system that performs these tasks and analysis.
[0024] With the GCS embodiments above, many different training and evaluation scenarios become available. As mentioned above, recording and storing the information provides data sets for training, and can also be reviewed offline. This allows instructors or other evaluators to evaluate the pilot's performance and the flight, regardless of whether the pilot is flying a simulation or in real-time, at a facility or out in the field.
[0025] In this manner, a training system allows a pilot and the system to track and monitor their stress levels during flights and provides real-time or near real-time feedback to allow the pilot to adjust. The embodiments focus on development of a unit that can be a critical part of emergency training. According to NASA data, 80% of Drone incidents occur due to Stress related issues with the Pilot.
[0026] All features disclosed in the specification, including the claims, abstract, and drawings, and all the steps in any method or process disclosed, may be combined in any combination, except combinations where at least some of such features and/or steps are mutually exclusive. Each feature disclosed in the specification, including the claims, abstract, and drawings, can be replaced by alternative features serving the same, equivalent, or similar purpose, unless expressly stated otherwise.
[0027] It will be appreciated that variants of the above-disclosed and other features and functions, or alternatives thereof, may be combined into many other different systems or applications. Various presently unforeseen or unanticipated alternatives, modifications, variations, or improvements therein may be subsequently made by those skilled in the art which are also intended to be encompassed by the embodiments.