SYSTEMS AND METHODS FOR MONITORING ACTIVITIES IN AN AVIATION ENVIRONMENT

20250131836 ยท 2025-04-24

    Inventors

    Cpc classification

    International classification

    Abstract

    The present invention is directed to systems and methods for monitoring activities in an aviation environment. The system includes at least two monitoring units, each including at least two types of sensors, wherein: the sensors are mounted at a plurality of locations in the aviation environment. The system further includes a processing system being configured to receive said information from the sensors, to process said information to monitor and make predictions, and to combine sensor information by applying data fusion. The system is further configured to compare sensor information with predetermined safety operation criteria, and to generate an alert signal. The method of the invention includes obtaining sensor information, receiving said information from the sensors at a processing system, processing said information, comparing the processed information with predetermined safety operation criteria, and generating an alert signal.

    Claims

    1. A system for monitoring activities in an aviation environment, the system including: at least two monitoring units, each monitoring unit including at least two types of sensors comprising a range sensor and a camera sensor, wherein: the sensors are adapted to obtain sensor information of at least two objects, including at least one runway and at least one aircraft, the monitoring units are mounted at a plurality of locations in the aviation environment, including at least one location at or near the runway; a processing system being configured to receive said information from the sensors and being further configured to process said information to monitor and make predictions in relation to the at least two objects, wherein: the processing system is configured to combine the range sensor information with the camera sensor information by applying data fusion to associate the sensor information with temporal information and spatial information; and applying data fusion includes processing the range sensor information and the camera sensor information acquired from the sensors at the plurality of locations in the aviation environment, including: applying a time-syncing process to the range and camera sensor information acquired at the plurality of locations; stitching or otherwise combining the range sensor information acquired at the plurality of locations; stitching or otherwise combining the camera sensor information acquired at the plurality of locations; and projecting or otherwise fusing the acquired range sensor information onto the acquired camera sensor information, or projecting or otherwise fusing the acquired camera sensor information onto the acquired range sensor information; the system being further configured to compare the temporally and spatially associated sensor information of the at least two objects with predetermined safety operation criteria, and to generate an alert signal when the compared information indicates a risk or likelihood of at least three different occurrence types from the following list of occurrence types: runway excursion; runway incursion; foreign object damage/debris; taxiing collision/near collision; unstable approach; wheels up landing; controlled flight into terrain; animal/bird strike.

    2. The system according to claim 1, wherein the system is configured to generate an alert signal when the compared information indicates a risk or likelihood of at least four different occurrence types from the list of occurrence types in claim 1.

    3. The system according to claim 1, wherein the system is configured to generate an alert signal when the compared information indicates a risk or likelihood of at least one occurrence from each of a first occurrence group of unsafe operation, and a second occurrence group of unsafe operation, wherein: the first occurrence group comprises one or more runway occurrence types on or near the runway, including one or more of, or any combination of: runway excursion, runway incursion, runway undershoot, depart/approach/land wrong runway, missed approach/go around, and/or rejected take off, and the second occurrence group comprises one or more aircraft control occurrence types, including one or more of, or any combination of: unstable approach, hard landing, ground strike, controlled flight into terrain, collision with terrain (near and at airport), collision/near collision (near and at runway), and/or ground proximity alerts/warnings.

    4. The system according to claim 3, wherein the system is further configured to generate an alert signal when the compared information indicates a risk or likelihood of at least one occurrence type from a third occurrence group comprising ground operation occurrence types; and the ground operation occurrence types comprise one or more of, or any combination of: foreign object damage/debris, jet blast/propeller/rotor wash, or taxiing collision/near collision.

    5. The system according to claim 4, wherein the system is further configured to generate an alert signal when the compared information indicates a risk or likelihood of at least one occurrence type from a fourth occurrence group comprising environment occurrence types, and the environment occurrence types comprise one or more of, or any combination of: icing, lightning strike, or animal/bird strike.

    6. The system according to claim 5, wherein the system is further configured to generate an alert signal when the compared information indicates a risk or likelihood of at least one occurrence type from a fifth occurrence group comprising infrastructure occurrences, including runway lighting occurrences or other infrastructure type occurrences.

    7. The system according to claim 1, wherein the range sensor comprises a LiDAR sensor and the processing system is configured to calculate range information of at least one object of the at least two objects by using sensor information from the LiDAR sensor.

    8. The system according to claim 7, wherein the processing system is configured to determine identity and/or classification information of at least one object of the at least two objects by using sensor information from the camera sensor and processing said sensor information using an artificial intelligence-based processing method.

    9. The system according to claim 8, wherein the processing system is configured to apply a deep- and/or machine-learning detection process to calculate the identity and/or classification information.

    10. The system according to claim 9, wherein the processing system is configured to associate the range information and the identity and/or classification information from the sensors to identify the at least one object in the field of view of the sensors.

    11. The system according to claim 10, wherein the processing system, configured to associate the at least one identified object with time information, thereby provides measurement and/or tracking at least one physical property of the at least one identified object over time.

    12. The system according to claim 11, wherein the processing system is configured to predict a physical property of the at least one identified object from tracked physical property information.

    13. The system according to claim 12, wherein the comparison of the information associated with the at least one identified object with the predetermined safety operation criteria includes measured physical property information and predicted physical property information from the at least one identified object.

    14. The system according to claim 11, wherein the measured and predicted physical property includes the aircraft's position, travel direction, velocity, acceleration, altitude and attitude, and other physical properties of aircraft of interest, and physical properties of other objects of interest including boundaries, markings, a centreline, a runway threshold, ground crew, a passenger, a ground vehicle, infrastructure and/or building structures.

    15. The system according to claim 1, wherein the system is configured to monitor the aircraft ground location and/or measure and/or calculate an estimate or prediction of the aircraft position or motion on the runway, aircraft position deviation from runway centreline, distance between aircraft and runway boundary and/or runway end, and a predicted time and/or position for runway excursion including veer-off.

    16. The system according to claim 1, wherein the system is configured to monitor an aircraft approach flight path and an aircraft landing configuration, to measure and/or calculate an estimate or prediction of acceptable deviation of measured flight path from an authorised or ideal flight path, and a likelihood of achieving safe touch-down or landing.

    17. The system according to claim 1, wherein the system is configured to monitor and/or track the aircraft location, and/or measure and/or calculate an estimate or prediction of the lift-off position, and a last safe stopping point along a take-off roll.

    18. The system according to claim 11, wherein the system is configured to receive and process additional information to assist and/or facilitate calculation of the at least two objects' physical properties, an estimation or prediction of their physical properties and/or the safe operation criteria.

    19-25. (canceled)

    26. A method for monitoring activities in an aviation environment, the method including the steps of: obtaining sensor information of at least two objects from at least two monitoring units, the objects including at least one runway and at least one aircraft, the at least two monitoring units being mounted at a plurality of locations in the aviation environment including at least one location at or near the runway; the monitoring units each housing at least two types of sensors, including a range sensor and a camera sensor, receiving said information from the sensors at a processing system being configured to process said information to monitor and make predictions in relation to said at least two objects; the processing system being configured to combine the range sensor information with the camera sensor information by associating the sensor information with temporal information and spatial information by applying data fusion, wherein: applying data fusion includes processing the range sensor information and the camera sensor information acquired from the sensors at the plurality of locations in the aviation environment, including; applying a time-syncing process to the range sensor information and the camera sensor information acquired at the plurality of locations; stitching or otherwise combining the range sensor information acquired at the plurality of locations; stitching or otherwise combining the camera sensor information acquired at the plurality of locations; and projecting or otherwise fusing the acquired range sensor information onto the acquired camera sensor information, or projecting or otherwise fusing the acquired camera sensor information onto the acquired range sensor information; comparing the processed information associated with the at least two objects with predetermined safety operation criteria, generating an alert signal when the compared information indicates a risk or likelihood of at least three different occurrence types from the following list of occurrence types: runway excursion; runway incursion; unstable approach; controlled flight into terrain; foreign object damage/debris; taxiing collision/near collision; animal/bird strike; wheels up landing.

    27. The method of claim 26, wherein the range sensor is a LiDAR sensor.

    28-29. (canceled)

    Description

    BRIEF DESCRIPTION OF THE DRAWINGS

    [0092] The present invention will now be described, by way of non-limiting example, with reference to the accompanying drawings in which:

    [0093] FIG. 1 is a functional diagram of a safety operation assessment system for monitoring activities in an aviation environment according to a preferred embodiment of the present invention;

    [0094] FIG. 2 is a schematic diagram illustrating a method for monitoring activities in an aviation environment according to a preferred embodiment of the present invention using the system of FIG. 1;

    [0095] FIG. 3 is an example flow-chart for the system and method of FIG. 1 for a particular occurrence type, runway excursion;

    [0096] FIGS. 4 to 6 are schematic diagrams illustrating runway excursion on landing, runway excursion on take-off and runway excursion veer-off respectively as illustrated in the flowchart of FIG. 3;

    [0097] FIG. 7 is an example flow-chart for the system and method of FIG. 1 for a particular set of occurrence types;

    [0098] FIG. 8 is a schematic diagram illustrating a particular set of occurrence types as illustrated in flow-chart of FIG. 7;

    [0099] FIG. 9 is an example flow-chart for the system and method of FIG. 1 for a particular set of occurrence types; and

    [0100] FIG. 10 is a schematic diagram illustrating a particular set of occurrence types as illustrated in flow-chart of FIG. 9.

    DETAILED DESCRIPTION OF THE INVENTION

    [0101] Preferred features of the present invention will now be described with particular reference to the accompanying drawings. However, it is to be understood that the features illustrated in and described with reference to the drawings are not to be construed as limiting on the scope of the invention.

    [0102] Referring now to FIGS. 1 to 10 there are illustrated safety operation assessment systems and methods for monitoring activities in an aviation environment according to preferred embodiments of the present invention.

    [0103] FIG. 1 illustrates a functional diagram of an exemplary system 2 within which the present invention may be embodied. The system 2 comprises a host service 4 (processing system) which is configured as described in greater detail below, in accordance with a preferred embodiment of the present invention, connected to a plurality of parties 16, 18, 20 over a network 6. The host service 4 is configured to facilitate engagement between at least one user 16, 18, 20, of the processing system 4 and one or more monitoring units 22 which can collect information from the aviation environment, particularly the aviation environment near and at airports. The users 16, 18, 20 are workers or companies that operate in the aviation environment, such as aircraft crew, ground crew, traffic control officers, emergency response teams and the like. The host service 4 are connectable via the network 6 to other third parties 24, for example fire attendance services or emergency government authorities or accident investigation agencies.

    [0104] The exemplary host service 4 comprises one or more host servers that are connected to a network 6, and therefore communicate via that network 6 via wired or wireless communication in a conventional manner as will be appreciated by those skilled in the art. The host servers are configured to store a variety of information collected from the users/units 16, 18, 20, 22 and 24.

    [0105] The host servers are also able to house multiple databases necessary for the operation of methods and systems of the present invention. The host servers comprise any of a number of servers known to those skilled in the art and are intended to be operably connected to the network so as to operable link to a computer system associated with the users 16, 18, 20 or third parties, 22 or 24. The host servers can be operated and supplied by a third party server providing service, or alternatively can be hosted locally by the processing system 4.

    [0106] The host server 4 typically includes a central processing unit (CPU) and/or at least one graphics processing unit (GPU) 8 or the like which includes one or more microprocessors, and memory 10, and storage medium 12 for housing one or more databases, operably connected to the CPU and/or GPU and/or the like. The memory 10 includes any combination of random-access memory (RAM) or read only memory (ROM), and the storage medium 12 comprises magnetic hard disk drives(s) and the like.

    [0107] The storage medium 12 is used for long term storage of program components as well as storage of data relating to the customers and their transactions. The central processing unit and/or graphics processing unit 8 which is associated with random access memory 10, is used for containing program instructions and transient data related to the operation of services provided by the host service 4. In particular, the memory 10 contains a body of instructions 14 for implementing at least part of a method for safety operation assessment in an aviation environment. The instructions 14 enable multiplatform deployment of the system 2, including on desktop computer, edge devices such as NVIDIA DRIVE or Jetson embedded platform. The instructions 14 also include instructions for providing a web-based user interface which enables users to remote access the system 2 from any client computer executing conventional web browser software.

    [0108] Each user 16, 18, 20, 22, 24 is able to receive communication from the host service 4 via the network 16 and is able to communicate with the host service 4 via the network 6. Each user 16, 18, 20, 22, 24 may access the network 6 by way of a smartphone, tablet, laptop or personal computer, or any other electronic device. The host service 4 may be provided with a dedicated software application which is run by the CPU and/or GPU and/or the like stored in the host servers. Once installed, the software application of the host service 4 provides an interface that enables the host service 4 to facilitate communication of information and/or alerts, including sensor information, raw or processed, to a predetermined user 16, 18, 20.

    [0109] In a preferred embodiment, the computing network 6 is the internet or a dedicated mobile or cellular network in combination with the internet, such as a GSM, CDMA, UTMS, WCDMA or LTE networks and the like. Other types of networks such as an intranet, an extranet, a virtual private network (VPN) and non-TCP/IP based networks are also envisaged.

    [0110] With reference to FIG. 2, method 100 has at least two sensors 26, 28, 30 for obtaining sensor information from at least one pre-determined location in the aviation environment. Each sensor is preferably of a different type to the other such that they obtain different sensor information, which advantageously complements each other's data acquisition capability. Preferably, each one of the at least two sensors 26, 28, 30 is housed in a plurality of monitoring units provided substantially equidistantly and/or strategically spaced about the aviation environment for the purposes of providing effective and efficient monitoring coverage of the operational aviation activity.

    [0111] Referring particularly to FIGS. 1, 2 and 4, there are provided about 10 or more monitoring units 22 which each include one of each of the least two sensors 26, 28, 30 and which are provided in multiple locations throughout the aviation environment near and at airport including runway, taxiway, apron, ramp areas, passenger boarding bridges, ground service vehicles, ground support vehicles, ground crew, airport building structures including gates. Monitoring units 22 are mounted on aircraft 16 and/or in locations on ground service vehicles and/or ground support vehicles 18 and equipment, ground personnel 20. Monitoring units 22 are configured and arranged so as to provide real-time, continuous and extensive views of a maximum space or volume near and at the airport (e.g. runway 40, taxiway 42, apron 44, ramp areas 46, runway threshold 48) in a variety of visibility or meteorological/environmental conditions. In particular the monitoring units 22 should also be configured to observe and monitor all, or a large proportion of, relevant aviation activities and operations near and at airport.

    [0112] In a preferred embodiment, one of the at least two sensor types is a Light Detection and Ranging (whose acronym is LiDAR) 26. LiDAR sensors 26 are particularly advantageous in extracting accurate range information of objects in its field of view. In a more preferred embodiment, another of the at least two sensor types is a light detector such as a camera 28, such as colour or infrared cameras or similar which can provide information about the at least one object of interest and/or their surrounding environment which enables object classification and tracking. Most preferably, each monitoring unit 22 has one of each of the LiDAR sensor 26 and a camera-type sensor 28 thereby advantageously providing range information of one or more objects and surrounding environment by LiDAR sensor, allowing accurate motion and position measurement; and providing visual information of one or more objects and surrounding environment by both LiDAR and camera-types sensors but primarily by the camera-type sensor, which facilitates accurate, precise and reliable object classification/recognition. Further, the two sensor types 26, 28 work together to provide the information in both normal and challenging light conditions such as fog, low light, sun glare smoke-filled, and the like, within the sensor's field of view and preferably up to 250 m from the monitoring unit. In particular, the LiDAR sensor may be adapted to work in foggy or rainy conditions by using 1505 nm wavelengths at higher power and/or using a Frequency-Modulated Continuous Wave radar or Full Waveform radar. Other sensor types 30 may be provided in the monitoring unit 22 and/or information acquired using other sensor types may be provided for the purposes of enhancing the system 2 or providing redundancies. Information may include meteorological, surface movement (incl. runway, taxiway, apron), aircraft data. The sensor types may include A-DBS, surface movement radar.

    TABLE-US-00001 TABLE 1 Pros and Cons of Example Sensor Information (including examples of preferred sensing properties) Cameras LiDAR Pros Real-time object detection High positioning accuracy <0.05 m Multiple object classes, e.g. aircraft High spatial resolution at a distance (complete aircraft, engine, landing gear), of up to 250 m, preferred spatial resolution ground vehicle, ground crew is <1.5 m High detection rate and accuracy, Range ~250 m both >95% May be adapted to work in low Simultaneous and 24/7 detection visibility conditions, e.g. fog, heavy rainfall Cons Relatively low object positioning Low object recognition capability accuracy, e.g. at a distance of 100 m, due to lack of visual detail and absence of accuracy is ~2 m colour Impacted by adverse light Object detection highly influenced conditions, e.g. strong shaded area, sun by reflective area of the target objects glare, fog, rain

    [0113] The sensors/monitoring units 26, 28, 30, 22 are also capable of producing and transmitting information from multiple locations to processing system 4 which is configured to receive said to process the information associated with the aviation activities in the operating aviation environment, particularly near and at airports. Preferably, the information is transmitted to the processing system 4 in a secured manner.

    [0114] The system 2 is configured to combine the information from the at least two types of sensors 26, 28, 30 acquired using at least one monitoring unit by associating the sensor information with time information, preferably by a processing system 4. The system 2 is also configured to combine the information from the at least two sensors 26, 28, 30 by associating the sensor information with spatial or distance or location information for example GPS coordinates or other positional information, range information and the like. The combination or fusing of the sensor information with time information may be obtained by time synchronisation or temporal calibration, while the combination or fusing of sensor information with spatial or distance or location information may be obtained by sensor calibration. At least one monitoring unit 22 can be employed to provide sensor information that can be fused into temporal and spatial data associated with objects in at least one predetermined location in the aviation environment, particularly near and at airports.

    [0115] More than one monitoring unit 22 is employed in areas, such as runways 40, apron 44, ramp areas 46, to monitor the same predetermined location, where the multiple monitoring units 22 are spaced apart thereby allowing combination of multiple sensor information associated with multiple monitoring units which is temporally synchronised and spatially calibrated as illustrated in FIGS. 3 to 10. In particular, the system 4 employs more numerous monitoring units 22 per unit area where the aviation environment has a large number of objects and a large activity volume, which might have high potential aviation safety risks. Further details will be provided in the following paragraphs.

    [0116] The processing system 4 is an artificial intelligence-based system which is configured to receive and process the sensor information to provide real-time sensing, recognition/classification and tracking of aircraft 16, ground personnel 20, ground vehicles 18 and other objects, recognition of operating environment, e.g. runway 40, taxiway 42, apron 44 and volume above these surfaces, and their features, e.g. runway boundary 49, marking 50, centreline 52, runway end 47, runway threshold 48, aircraft engine 51, aircraft landing gear 53, object motion and position estimation. The sensor information may be fused, i.e. temporally synchronised and/or spatially calibrated once received by the processing system 4 or alternatively it may be fused beforehand.

    [0117] The terms artificial intelligence and intelligent algorithms are used herein to refer to and encompass systems of data processing and analysis that are conducted by computers capable of harvesting large amounts of possible input data, including images and other information from monitoring and sensing devices, that may be processed, analysed, and categorized based on a set of rules and then may be communicated so that appropriate action may be taken, whether automatically by a system receiving the processed and analysed data or manually by at least one human operator such as Air Traffic Control officer, pilot and emergency response team.

    [0118] Referring to FIGS. 1 and 2, the processing system 4 is further configured to process the sensor information including the following example steps of a method 100 and data processing step 104 for safe operation assessment in an aviation environment which is summarised in Table 2, below.

    TABLE-US-00002 TABLE 2 Example Data Processing Stages Stage 1: Data acquisition, preparation and fusion A. Sensor calibration Extrinsic, intrinsic and distortion parameters of sensors, i.e. LiDAR and camera This may include individual and/or cross calibration Quantify sensor errors B. Time synchronisation Between LiDAR and camera data acquired at various locations C. Data fusion Project LiDAR data, i.e. point cloud, on image or vice versa Data stitching between sensors data acquired at various locations for both LiDAR and camera, for example, LiDAR data stitching/registration using Iterative Closest Point (ICP), normal distributions transform (NDT), phase correlation, coherent point drift (CPD). Stage 2: Segmentation A. Ground plane segmentation (LiDAR points) Separate foreground (e.g. aircraft) and background (e.g. runway) objects using ground plane estimation B. Cluster 3-D points (LiDAR points) Apply output of Stage 3 Step A to output in Stage 2 Step A to form an object using a 3-D points grouping/clustering process Stage 3: Object detection/Identification A. Camera frames (2-D images) Train, test, optimise and apply deep learning-based object detection models, e.g. You Only Look Once (YOLO), Fully Convolutional One-Stage (FCOS) B. 3-D space (LiDAR points) e.g. use spin image method for 3-D object recognition e.g. use PointSeg network C. Detection confidence score (2-D images and LiDAR points combined) Integrate confidence score results from Steps 3.A and 3.B Stage 4: Multiple Object Tracking and motion estimation Associate moving objects in the current frame and the previous frame. A. Camera frames (2-D images) e.g. use Kalman filter method B. 3-D space (LiDAR points) e.g. use segment matching based method e.g. joint probabilistic data association (JPDA) tracker Stage 5: Safe operation assessment Use data acquired and processed from Stages 1 to 4, measure, calculate and predict: moving object's position, travel direction, velocity, acceleration distance between aircraft of interest and object of interest, e.g. runway centreline, boundary, other aircraft Assess/calculate, and predict/determine/decide & generate alert when deviation from and/or violation of safety operation criteria is detected.

    [0119] With reference to Table 2 and FIG. 2, the information/data is received from the at least two sensors 26, 28, 30 or at least one monitoring unit 22 in step 102 and is prepared for being received by the processing system 4 in step 104. Next, in step 104 generally speaking, the processing system 4 processes the sensor information.

    [0120] In this particular example, see step 104 exemplified by table 2, the system 4 is configured to receive sensor information from the camera 28 and LiDAR 26 and to combine the two type of sensors' information by data fusion methods, including by sensor calibration and/or time-syncing.

    [0121] Preferably, the data fusion, and preparation of the data therefor, includes acquisition of extrinsic, intrinsic and distortion parameters of sensors (i.e. LiDAR and camera), followed by quantification of sensor errors.

    [0122] Preferably, time synchronisation may be achieved through the use of internal/external timer source(s) that are coupled to with the sensors, and the read and comparison of timestamps that are associated with individual image and point cloud data using the processing system.

    [0123] Preferably, the LiDAR information, a 3-D point cloud of the objects within the aviation environment, is projected on the camera image or vice versa.

    [0124] Preferably, the LiDAR information, acquired from multiple LiDAR sensors that are located at various locations, is registered/stitched/fused using algorithms such as Iterative Closest Point (ICP), normal-distributions transform (NDT), phase correlation, coherent point drift (CPD).

    [0125] Preferably, the image information, acquired from multiple cameras that are located at various locations, is registered/stitched/fused using algorithms such as feature based image registration. The abovementioned operations may be incorporated in alternative examples or embodiments of the present invention.

    [0126] It will be understood that the person skilled in the art would be able to conduct data fusion (e.g. sensor calibration, time-syncing) by a variety of methods or algorithms.

    [0127] In the next step Stage 2 in the example described in Table 2, the system 2, and more preferably the processing system 4, is configured to process the sensor information to separate the foreground from the background via ground plane segmentation process(es). In this example, 3-D LiDAR point cloud obtained in the previous steps is used to separate foreground objects, such as aircraft or support ground-based vehicles, from background objects i.e. runway. In particular, the processing system 4 can perform the separation or ground plane segmentation by techniques such as ground plane estimation, however it is expected that other known techniques could be utilised.

    [0128] Next the processing system 4 is then configured to form at least one object from 3-D LiDAR point cloud. In this example, the 3-D point cloud object is formed by the result of the combination of received outputs produced by separation of the foreground and background in the previous step (Stage 2 Step A in Table 2) and object detected and classified from the camera image which is processed in the Stage 3 Step A in Table 2. Preferably the object formed by combination is formed by a 3-D points grouping or clustering process thereby forming a 3-D space although it would be understood that other processes or techniques could be equally employed. The Stage 3 Step A, camera image processing step, is independently processed to the Stage 2 steps and therefore can be performed temporally before or in parallel with Stage 2 Step A such that the results of Stage 3 Step A is available and ready for use before the commencement of Stage 2 Step B. Results of Stage 3 Step A is an input to Stage 2 Step B.

    [0129] In the next processing step, Stage 3, as illustrated in Table 2, the processing system 4 is configured to detect and/or identify and classify objects in the aviation environment. In the example Stage 3 summarised in Table 2, the processing system 4 can first process the camera sensor information received from the camera images/video frames to detect and/or identify the objects. In particular, the artificial intelligence-based data processing system 4 employs machine- or deep-learning-based object detection and/or classification models, which are trained, validated, verified and optimised for detection and classification of objects involved in aviation activities in an aviation environment near and at airport. The object detection and/or classification models that can be utilised include You Only Look Once (YOLO) or Fully Convolutional One-Stage (FCOS) models although it is expected that other artificial intelligence-based models could equally be used instead for similar effect.

    [0130] In data processing Stage 2 Step B, the processing system 4 can also process the LiDAR sensor information which has been processed to form a cluster 3-D points in Stage 2 Step B, for object identification and/or recognition. Example techniques for the 3-D object recognition in the 3-D space can include the spin image method or the PointSeg network although other known methods could be utilised.

    [0131] In the next processing step, Stage 3 Step C, the processing system 4 can then combine the processed camera sensor information and processed LiDAR sensor information which can result in a detection confidence score which can be associated with the sensor information. The use of detection confidence score enhances the detection and classification accuracy by reducing false detections and by increasing detection rate. For example, for reducing false detection, there are two aircraft have similar configurations and features but are different in size, i.e. both are configured with a cylindrical fuselage with two jet engines, and one aircraft is 30 metres long whereas another aircraft is 60 metre long. If the larger aircraft is located closer than the smaller aircraft to the camera, information acquired from the camera and subsequently processed by the processing system might not be able to accurately differentiate the size difference between the two aircraft. The information about these two aircraft acquired from LiDAR, on the other hand, can provide accurate size information and location information of these two different types of aircraft regardless of the difference in distance between the aircraft and the LiDAR sensor. For example, for increasing detection rate, while LiDAR information provides high positioning accuracy of 0.05 metres, the spatial resolution of 1.5 meters at a distance of 200 metres may be sufficient to detect and identify an aircraft with a length of 30 metres, but it may not be able to detect and identify objects with dimensions below 1.5 metres such as some ground equipment, e.g. tow bar 45, ground crew, cargo/baggage cart. By combining these two types of information acquired with camera and LiDAR sensors, the detection and classification accuracy may therefore be enhanced by reducing the effects of lack of range information from camera information and by reducing the effects of lack of visual detail and absence of colour from 3-D LiDAR point cloud.

    [0132] Once the system 2 has detected, identified and/or classified the objects in the aviation environment near and at airports, the system 2 is then configured to associate the motion of at least one object, preferably multiple objects, over time as exemplified in the example Stage 4 of Table 2. Further the system 2 is also configured to provide an estimation of the motion of the object(s). For the purposes of object tracking and motion estimation, the system 2 is configured to associate moving objects in one information acquisition and at least one other subsequent information acquisition. One information acquisition refers to one camera/video frame and one LiDAR frame or its equivalent, which are temporally-synchronised and spatially calibrated. The system 2, particularly the processing system 4, is configured to process the sensor information from the 2-D camera/video images and/or the LiDAR point clouds from the 3-D space to associate sensor information from each sensor from one information acquisition (i.e. camera/video frame and/or LiDAR point clouds) to a subsequent or previous information acquisition. Preferably, the processing system 4 is able to process sensor information associated with at least two sequential camera/video frames at a particular moment when information acquisitions are received by the data processing system 4 continuously over time. In a particularly preferred embodiment, the processing system 4 employs the Kalman filter method to process the 2-D camera/video images, and the segment matching based method or joint probabilistic data association (JPDA) tracker to process the 3-D space data (LiDAR point clouds). It would be understood however that other models or methods to predict the physical properties of the objects' predicted physical properties could be substituted for the ones named above.

    [0133] In a final processing stage (stage 5 of the example processing method in Table 2), the system 2 is configured to combine the outputs of the processed sensor information from the previous steps/stages to measure, calculate and provide an output of the estimation or prediction of one or more objects' physical properties such as position, acceleration, speed and/or travel direction of any object(s) motion. Furthermore, the system 2 is configured to compare one predicted object's physical properties to another, for example a distance or predicted distance between aircraft 16 and another object of interest, i.e. runway centreline 52, boundary 49, runway threshold 48, other aircraft 16, and to output information which is associated with these properties of the compared objects. The system 2 is able to assess the properties of the compared objects with predetermined safe operation criteria and to generate an alert (in step 106, see FIG. 2) when the system 2 has determined that the predetermined safe operation criteria may or has been violated or otherwise deviated therefrom, and to provide suggested corrective or evasive actions to a number of users in the aviation environment, particularly near and at airports.

    [0134] Although the examples described herein refer to an aviation environment, particularly near and at airports, the system 2 can be utilised in a number of other environments requiring monitoring of multiple moving and static objects within an environment such as industrial environments, such as maritime operations, road/autonomous driving operations, mining operations, industrial plants, logistics centres, manufacturing factories, aviation operations that are not near and at airports, space operations and the like.

    [0135] Details of the various predetermined safe operation criteria are provided in the following paragraphs and in particular from Tables 3 and 4.

    TABLE-US-00003 TABLE 3 Example of occurrence types, Detection and Tracking Multiple Objects data processing capability and Safe Operation Criteria. Functions/Occurrence Detection and Tracking Multiple Objects (DATMO) Capability Types Safe Operation Criteria RUNWAY Object detection A1. Runway Moving object: aircraft, ground vehicle, ground crew excursion Static object: runway, taxiway, apron (surface, markings, A2. Runway boundary, centreline) incursion Object classification: aircraft type, airline A3. Runway Object tracking: aircraft, ground vehicle, ground crew undershoots Measure, calculate and predict: A4. Rejected take-off Moving object's position, travel direction, velocity, A5. Depart/Approach/ acceleration, altitude and attitude (if airborne) Land Wrong Runway Distance between aircraft of interest and object of interest, e.g. A6. Missed approach/ runway centreline, boundary, other aircraft go-around Coverage: On and near runway (e.g. a volume 200 m higher and 500 m A7. Runway events - wider than runway surface), taxiways and apron/stand/gate Other Calculate, and predict/determine/decide & alter: Runway distance remaining & likelihood of runway excursion and rejected take-off Unauthorized entry of aircraft, ground vehicle and crew, vehicle & likelihood of runway incursion On approach, aircraft altitude (above ground), vertical speed, distance between aircraft and runway boundary, and deviation from desired approach profile & likelihood of runway undershoot After take-off or on approach, compare actual flight path with designated/expected flight path & likelihood of depart/approach/ land wrong runway, missed approach/go-around GROUND Object detection OPERATIONS Moving object: aircraft, ground vehicle, ground crew/person, B1. Foreign object cargo, fuel damage/debris Static object: runway, taxiway, apron (surface, markings, B2. Objects falling boundary, centreline), cargo, fuel, foreign object/debris/aircraft from aircraft parts B3. Taxiing Aircraft features: engine, propeller, rotor, auxiliary power unit, collision/Near collision fuel, damage, ice B4. Jet Object classification: aircraft type, airline, foreign object/debris/ blast/Prop/Rotor wash aircraft parts B5. Ground handling Object tracking: aircraft, ground vehicle, ground crew, cargo B6. Ground Measure, calculate and predict operations - Other Moving object's position, travel direction, velocity, B7. Interference with acceleration, altitude and attitude (if airborne) aircraft from ground Distance between aircraft of interest and object of interest, e.g. B8. Dangerous goods runway and taxiway surface, markings, centreline, boundary, other B9. Loading related aircraft, ground vehicle, ground crew B10. Aircraft loading - Coverage: On and near runway (e.g. a volume 200 m higher and 500 m Other wider than runway surface), taxiways and apron/stand/gate B11. Fuel leaking of Calculate, and predict/determine/decide & alter venting Foreign object, debris, aircraft parts and likelihood of foreign B12. Auxiliary power object damage/debris, objects falling from aircraft unit Position, distance, velocity and acceleration, and travel B13. Engine failure or direction between aircraft and object of interest (aircraft, ground malfunction vehicle, ground crew, infrastructure) & likelihood of collision, ground B14. Fuselage/Wings/ handling, loading related, interference with aircraft from ground Empennage occurrences B15. Anti-ice Distance between aircraft engine/propeller/rotor and object of protection interest (aircraft, ground vehicle, ground crew, infrastructure) & B16. Security related likelihood of jet blast/prop/rotor wash Detect and monitor likelihood of fuel leaking or venting, dangerous goods, auxiliary power unit, aircraft damage, engine failure or malfunction, ice on aircraft, security related occurrence AIRCRAFT Object detection CONTROL Moving object: aircraft, ground vehicle, ground crew C1. Hard landing Static object: runway, taxiway, apron (surface, markings, C2. Ground strike boundary, centreline) C3 Wire strike Aircraft features: flap, landing gear C4. Loss of control Object classification: aircraft type, airline C5. Unstable Object tracking: aircraft, ground vehicle, ground crew approach Measure, calculate and predict C6. Wheels up Moving object's position, travel direction, velocity, landing acceleration, altitude and attitude (if airborne) C7. Landing gear/ Distance between aircraft of interest and object of interest, e.g. Indication runway centreline, boundary, other aircraft C8. Incorrect Coverage: On and near runway (e.g. a volume 400 m higher and configuration 5000 m longer than runway surface) C9. Ground Calculate, and predict/determine/decide & alter proximity alerts/ On approach, aircraft vertical deceleration exceeds the limit set warnings in the aircraft's operations manual or damage occurs during the C10. Flight below landing & likelihood of hard landing LSALT On approach, distance between aircraft fuselage and runway C11. Lost/unsure of surface, fumes or spark due to contact between aircraft and ground position surface & likelihood of ground strike C12. Collision with After take-off or on approach, compare actual flight path with terrain (near and at designated/expected flight path & likelihood of ground proximity, airport) loss of separation, collision (near and at runway), loss of control, C13. Collision/Near flight below LSALT, lost/unsure of position, wire strike, collision (near and at controlled flight into terrain, airspace infringement, failure to pass runway) traffic, information/procedural error, ANSP operational error - C14. Controlled flight other, operational non-compliance, airspace - other into terrain (CFIT) On approach, aircraft's deviation from the aircraft approach C15. Loss of profile parameters stipulated in a company's standard operating separation procedures (SOPs) & likelihood of unstable approach C16. Loss of On approach, whether the aircraft has timely and correctly lowered separation assurance landing gear and configured flap & likelihood of wheels up landing, (LOSA) landing gear/indication, incorrect configuration C17. Aircraft separation - issues C18. Airspace infringement C19. Operational non- compliance C20. Airspace - other ENVIRONMENT Object detection D1. Animal strike Moving object: aircraft, ground vehicle, ground crew, wildlife, D2. Bird strike remotely piloted aircraft D3. Wildlife - Other Static object: runway, taxiway, apron (surface, markings, D4. Near encounter boundary, centreline, lighting), wildlife, remotely piloted aircraft with remotely piloted Aircraft features: falling parts, aircraft damage, fire, fume, aircraft smoke, icing D5. Turbulence/ Object classification: aircraft type, airline, wildlife, remotely piloted windshear/microburst aircraft D6. Unforecast Object tracking: aircraft, ground vehicle, ground crew weather Measure, calculate and predict D7. Emergency Moving object's position, travel direction, velocity, evacuation acceleration, altitude and attitude (if airborne) D8. Fire Distance between aircraft of interest and object of interest, e.g. D9. Fumes remotely piloted aircraft, wildlife, foreign object/debris/aircraft D10. Smoke parts D11. Icing Coverage: On and near runway (e.g. a volume 400 m higher and D12. Lightning strike 5000 m longer than runway surface), taxiways and apron/stand/gate Calculate, and predict/determine/decide & alter Wildlife, remotely piloted aircraft & likelihood of animal strike, bird strike, wildlife - other, near encounter with remotely piloted aircraft Turbulence/windshear/microburst, unforecast weather and likelihood of turbulence/windshear/microburst, unforecast weather Likelihood of emergency evacuation Fire, fume, smoke & likelihood fire, fume, smoke, icing or lightning strike INFRASTRUCTURE Object detection E1. Rada/ Moving object: aircraft, ground vehicle, ground crew surveillance Static object: runway, taxiway, apron (surface, markings, E2. Runway lighting boundary, centreline), and infrastructures E3. Infrastructure - Aircraft and environment features: fire, fume, smoke other Infrastructure may include, one or more buildings, gates, hangars, light polls, and/or fences Object classification: aircraft type, airline, fire, fume, smoke Object tracking: aircraft, ground vehicle, ground crew Measure, calculate and predict Moving object's position, travel direction, velocity, acceleration, altitude and attitude (if airborne) Distance between aircraft of interest and object of interest, e.g. fire, fume, smoke Coverage: On and near runway (e.g. a volume 400 m higher and 5000 m longer than runway surface), taxiways and apron/stand/gate Calculate, and predict/determine/decide & alter: Any faults or deficiencies associated with radar/surveillance, runway lighting, other infrastructure & likelihood of radar/surveillance, runway lighting and other infrastructure occurrences

    TABLE-US-00004 TABLE 4 Example safe/unsafe operation criteria and assessment method. Safe/unsafe operation criteria Safe operation assessment method A1. Runway excursion A1.1. Veer off side of runway An aircraft that veers off the side of the runway Predict likelihood of runway excursion or overruns the runway threshold. Excursion by monitoring deviation of aircraft track from occurrences occur during take-off or landing runway centreline, i.e. tendency to deviate from only, and may be either intentional or runway centreline. unintentional. Predict likelihood of runway excursion A1.1. Veer off side of runway by monitoring distance between aircraft landing Deviation (i.e. distance) of aircraft from gears/fuselage/wingtip and runway side the centreline of the runway should not be boundary. significant, e.g. for a runway with a width of 60 A.1.2. Overrun runway threshold meters and an aircraft with a wing span of 60 Monitor aircraft position and motion meters, the centreline of the aircraft should not (e.g. travel direction, velocity and acceleration) deviate more than 15 meters from runway on runway. centreline. Monitor distance between aircraft and Distance between aircraft landing gear runway threshold, i.e. runway distance and runway side boundary should not be remaining. significant, e.g. for a runway with a width of 60 Predict likelihood of runway excursion meters and an aircraft with a wing span of 60 by calculating lift-off position (take-off case) or meters, the distance between aircraft landing position that ground speed is lowered to a gear and runway side boundary should not be certain value (landing case) based on runway less than 15 meters. distance remaining, and present and predicted A1.2. Overrun runway threshold aircraft motion. Aircraft should not overrun runway Calculations should also be based on threshold. aircraft breaking capability (e.g. specific aircraft During take-off, aircraft should type's spoiler, tyre break, engine anti-thruster accelerate to become airborne before overruns performances) and runway conditions (e.g. dry, threshold, e.g. for a runway with a length of wet, surface material type) and other 3000 meters, under certain weather and runway contributing factors. surface conditions, for a particular aircraft type Alerts and weight, the predicted lift off position is If likelihood of runway excursion within 1600 meters from start position with a lift off next 20 seconds is high and persist for more speed (VLOF) of 120 knots, and runway distance than 2 seconds, generate alerts. remaining should be 1400 meters. During landing, after touching down, aircraft speed should become low enough to ensure a safe stop before the end of the runway and/or aircraft can safety exit runway. A2. Runway incursion Detect and track aircraft, vehicle or The incorrect presence of an aircraft, vehicle or person on the protected area of a surface person on the protected area of a surface designated for the landing and take-off of designated for the landing and take-off of aircraft. aircraft. Predict likelihood of runway incursion anything within the confines of the by checking authorisation for detected aircraft, runway strip, irrespective of having an vehicle or person. appropriate clearance, which hinders the Alerts operation of an arriving or departing aircraft; or If likelihood of runway incursion within an aircraft, vehicle or person entering next 20 seconds is high and persists for more the confines of the flight strip without a than 2 seconds, generate alerts. clearance to do so, regardless of other aircraft operations. A3. Runway undershoots Monitor aircraft flight path on approach, Any aircraft attempting a landing and touches e.g. lateral and vertical profile, airspeed and down prior to the threshold. Any occurrence bank angle, altitude, vertical speed. where an aircraft touches down short of the Predict likelihood of runway approved designated landing area of the runway - undershoots by calculating touch down generally relates to a misjudgement by a pilot point/impact point based on present and during the approach phase. For example, predicted aircraft flight path on approach. aircraft that come into contact with vegetation Alerts or a fixed object (fence line, powerline, etc). If likelihood of runway undershoots and continues the approach. within next 20 seconds is high and persists for more than 2 seconds, generate alerts. A4. Rejected take-off Monitor aircraft take off performance, Any circumstance by which aircraft including aircraft motion along runway (e.g. discontinues the take-off after commencement velocity, acceleration), aircraft configuration of the take-off roll. The situation which follows (e.g. flap, wing surface de-ice), aircraft damage when it is decided to stop an aircraft during the (e.g. engine, fuselage), runway conditions (e.g. take off roll and may be initiated by flight crew unauthorised aircraft/ground/vehicle/person, or air traffic control officer. foreign object damage/debris, surface For example, an aircraft should discontinue the conditions), etc. take-off roll if any circumstance that may Predict likelihood of rejected take-off endanger aviation safety is identified, e.g. by monitoring aircraft take off performance engine malfunction/failure, runway incursion. based on aircraft motion, aircraft configuration, aircraft damage and runway conditions. Alerts If likelihood of rejected take-off within next 20 seconds is high and persists for more than 2 seconds, generate alerts A5. Depart/Approach/Land Wrong Monitor track (ground movement, flight Runway path) of an aircraft during takes off, landing, Depart/Approach/land wrong runway includes attempts to land from final approach, operates occurrences where a pilot unintentionally: in the circuit. approaches, takes off from, or lands on Predict likelihood of Depart/Approach/ a runway other than that intended or authorised Land Wrong Runway by assessing deviation of by air traffic control officer monitored track from intended or authorised approaches, takes off from, or lands on track. a closed runway Alerts approaches, takes off from, or lands on If likelihood of Depart/Approach/Land a taxiway Wrong Runway within next 20 seconds is high approaches and/or lands on a roadway and persists for more than 2 seconds, generate in the vicinity of an aerodrome alerts. A6. Missed approach/go-around Monitor aircraft activities on approach Any circumstance in which the aircraft that might stipulate a missed approach discontinues its approach to land. A missed procedure. approach procedure is the procedure to be Predict likelihood of a missed approach followed if an approach can no longer be procedure should be carried out. continued based on the flight crew or air traffic Alerts control officer assessment that the approach has If likelihood of missed approach/go-around been compromised. The missed approach should be carried out procedure takes into account de-confliction within next 20 seconds is high and from ground obstacles and from other air traffic persists for more than 2 seconds, generate flying instrument procedures in the airfield alerts. vicinity. Reasons for discontinuing an approach include the following: required visual references have not been established by Decision Altitude/Height (DA/H) or Minimum Descent Altitude/Height (MDA/H) or is acquired but is subsequently lost approach is, or has become unstabilised aircraft is not positioned so as to allow a controlled touch down within the designated runway touchdown zone with a consequent risk of aircraft damage with or without a Runway Excursion if the attempt is continued runway is obstructed, e.g. ground obstacles and from other air traffic flying instrument procedures in the airfield vicinity a landing clearance has not been received or is issued and later cancelled A7. Runway events - Other Monitor runway, taxiway, apron, ramp Runway event occurrences not specifically areas. covered elsewhere. Runway hazards do not Monitor moving and static objects belong to Al-A6. Add and assess in accordance (aircraft/terrain/ground with local operating conditions. vehicle/person/object/terrain). B1. Foreign object damage/debris Detect and monitor foreign object, Any loose objects on an aerodrome or in an debris or any loose objects on aerodrome or in aircraft that have caused, or have the potential an aircraft. to cause, damage to an aircraft. Monitor distance between aircraft and The operation of an aircraft has been affected - foreign object, debris or any loose objects. i.e. causes damage, aircraft passes over object, Predict likelihood of foreign object, rejected take off or missed approach. debris or any loose objects adversely affecting on the operating aircraft. Alerts If likelihood of foreign object damage/ debris is high and persists for more than 5 seconds, generate alerts. B2. Objects falling from aircraft Detect and monitor objects falling from The unintentional loss of an aircraft aircraft. component or object inside or on the aircraft Monitor distance between aircraft and that falls to the ground or detaches from the objects falling from aircraft. aircraft during normal flight operations. This Predict effects of objects falling from includes: aircraft on operating aircraft at airport. aerials Alerts lights If objects falling from aircraft is panels adversely affecting operating aircraft near and external loads (helicopter) at airport, generate alerts. wheels If objects falling from aircraft is detected, inform relevant stakeholder, airport engineering team, to inspect/remove. B3. Taxiing collision/Near collision Monitor aircraft tracks, and its distance An aircraft collides, or has a near collision, with to other aircraft/terrain/ground another aircraft, terrain, person, ground vehicle vehicle/person/object. or object on the ground or on water during taxi. Monitor awareness of one aircraft to the An aircraft collides, or has a near collision, with other aircraft/terrain/ground another aircraft, terrain, person, ground vehicle vehicle/person/object, avoiding plan and/or or object on the ground or on water during taxi. action. Monitor damage to the fuselage, wings, or empennage Predict likelihood of collision by monitoring distance between aircraft and other aircraft/terrain/ground vehicle/person/object and present and predicted aircraft tracks (e.g. path, travel direction, velocity). Alerts If likelihood of collision within next 20 seconds is high and persists for more than 2 seconds, generate alerts. B4. Jet blast/Prop/Rotor wash Monitor ground-running aircraft Any air disturbance from a ground-running propeller, rotor or jet engine. aircraft propeller, rotor or jet engine that has Monitor other aircraft/terrain/ground caused, or has the potential to cause, injury or vehicle/object/person near the ground-running damage to property. aircraft propeller, rotor or jet engine. Any air disturbance from a ground-running Monitor distance between ground- aircraft propeller, rotor or jet engine should not running aircraft propeller, rotor or jet engine cause, or has the potential to cause, injury or and other aircraft/terrain/ground damage to property (e.g. other vehicle/object/person. aircraft/terrain/ground vehicle/object) or person. Predict likelihood of jet blast/prop/rotor wash by monitoring distance between aircraft and other aircraft/terrain/ground vehicle/person/object, and present and predicted aircraft tracks. Alerts If likelihood of jet blast/prop/rotor wash within next 20 seconds is high and persists for more than 2 seconds, generate alerts. B5. Ground handling Monitor ground handling and aircraft Any ground handling and aircraft servicing that servicing by detecting, tracking and monitoring caused, or has the potential to cause injury or aircraft, vehicles, persons, equipment. damage to a stationary aircraft. Monitor compliance of ground handling Ground handling relate specifically to ramp and aircraft servicing with defined procedures, operations - i.e. engineering, aircraft loading, e.g. fuel spillage, ramp operations - i.e. catering and refueling services, etc. This can engineering, aircraft loading, catering and take place on the land or water, and include refuelling services, pushback procedures or operations on ships, oil rigs, and similar other engineering related occurrence. platforms. This includes: Monitor distance between aircraft, vehicles colliding with a stationary vehicles, persons, equipment, e.g. vehicles aircraft colliding with a stationary aircraft. fuel spillages Predict likelihood of ground handling pushback procedures or other by monitoring distance between aircraft, engineering related occurrence vehicles, persons, equipment, and present and predicted tracks, and compliance with defined procedures. Alerts If likelihood of ground handling within next 20 seconds is high and persists for more than 2 seconds, generate alerts. B6. Ground operations - Other Monitor runway, taxiway, apron, ramp Ground operation occurrences not specifically areas. covered elsewhere. Monitor moving and static objects Runway hazards do not belong to B1-B5. (aircraft/terrain/ground Add and assess in accordance with local vehicle/person/object/terrain). operating conditions. B7. Interference with aircraft from ground Detect and monitor interference sources near and at airport. Any ground based activity that interferes with Predict likelihood of interference to the operation of an aircraft. Ground based aircraft from ground based on interference occurrence types: distance/proximity between aircraft and laser/Spotlight interference source(s) near and at airport, and model aircraft effects of interference(s). radio frequency interference Alerts weather balloons If likelihood of interference to aircraft yacht masts from ground within next 20 seconds is high and persists for more than 5 seconds, generate alerts. B8. Dangerous goods Monitor and classify goods, including The carriage of dangerous goods in undeclared dangerous goods. contravention of Commonwealth, State or Monitor and detect spills, incorrect Territory law. packing/stowing. Dangerous goods occurrences include situations Monitor and detect aircraft trimming in which: and weight and balance issues undeclared dangerous goods are Predict likelihood of dangerous goods discovered and loading related occurrences by monitoring dangerous goods have spilled goods, loading procedures, spills, etc. dangerous goods are incorrectly packed Alerts or stowed If likelihood of dangerous goods and loading B9. Loading related related occurrences is high and persists for more The incorrect loading of an aircraft that has the than 2 seconds, generate alerts. potential to adversely affect any of the following: the aircraft's weight; balance; structural integrity; performance; flight characteristics. Freight issues occurrences include: incorrect load sheets; freight shifting in flight; unrestrained or inadequately restrained freight; spillages in a freight hold (other than dangerous goods); an incorrectly trimmed aircraft; weight & balance issues B10. Aircraft loading - Other Aircraft loading occurrences not specifically covered elsewhere that do not belong to B8-B9. Add and assess in accordance with local operating conditions. B11. Fuel leaking of venting Similar to B1, B2, E2 Relates specifically to the unplanned loss of fuel from a fuel tank or fuel system. B12. Auxiliary power unit Similar to D8, D9, D10 Any mechanical failure of the APU i.e. APU fires, fumes and smoke events where the APU was identified as the source B13. Engine failure or malfunction An engine malfunction that results in a total engine failure, a loss of engine power or is rough running includes: Engine fires, fumes and smokes where the engine was identified as the source A rough running engine (coughing, spluttering, etc) B14. Fuselage/Wings/Empennage Similar to B1, B2, E2 Damage to the fuselage, wings, or empennage Security related incidents on their own are to be not caused through collision or ground recorded as an Event. A scheduled report of contact. Any damage to the fuselage, wings, or all reported security related matters are sent to empennage that involve: cracks, creases, dents The Office of Transport Security on a weekly B15 Anti-ice occurrence types include: pitot basis. heat, deice boots, carburettor heat, nacelle/engine anti-ice B16. Security related When aviation security has been, or is likely to have been, compromised includes situations involving: weapons or prohibited items being taken onto an aircraft. the discovery of unidentified or suspicious objects on an aircraft. attempted unlawful interference, such as sabotage, hijack, vandalism etc. unapproved airside entry of persons or vehicles C1. Hard landing The vertical deceleration limit for the aircraft set out in the aircraft's On approach, monitor and calculate operations manual is exceeded or damage aircraft vertical deceleration occurs during the landing. Monitor damage occurs during the The vertical deceleration limit for the landing aircraft set out in the aircraft's operations Predict likelihood of hard landing by manual should not be exceeded and/or damage comparing monitored and calculated aircraft should not occur during the landing. vertical deceleration with the limit set in the aircraft's operations manual Alerts If likelihood of hard landing within next 20 seconds prior to touch down is high and persists for more than 2 seconds, generate alerts. C2 Ground strike Monitor distance/clearance between When part of the aircraft drags on, or strikes, aircraft (e.g. rotor, propeller, engine pod, the ground or water. Ground strike includes wingtip, tail) and ground or water/wire. situations where an aircraft is in the take-off or Predict likelihood of ground strike by landing phase a (including a hover taxi for monitoring distance/clearance between aircraft helicopters) in which: and ground or water/wire and comparing that a rotor or propeller makes contact with value with a defined/recommended safe the ground distance. an engine pod, wingtip, or tail contacts Alerts the ground If likelihood of ground/wire strike within next C3. Wire strike 20 seconds is high and persists for more than 2 When an aircraft strikes a wire, such as a seconds, generate alerts. powerline, telephone wire, or guy wire, during normal operations. C3's criteria is similar to C2's C4. Loss of control Monitor and track aircraft path on When control of the aircraft is lost or there are ground and airborne near and at airport. significant difficulties controlling the aircraft Predict likelihood of a loss of control either airborne or on the ground. occurrence based on extent of deviation of Loss of control occurrences include: present and predicted path from normal path. an unintentional ground loop of an Alerts aircraft1 If likelihood of loss of control within next 20 unintentional departure from normal seconds is high and persists for more than 2 flight necessitating recovery action or resulting seconds, generate alerts. in a terrain collision helicopter dynamic rollover C5. Unstable approach As a general guide, when an aircraft is A continued approach and/or landing in on approach and within 1,000 feet above the contravention of the operator SOP relating to aerodrome, monitor their stable approach criteria. track/localiser deviation An aircraft should not continue to land from an descent rate approach where there is sufficient evidence of a altitude significant deviation from the aircraft approach bank angle profile parameters stipulated in a company's alignment with runway centreline standard operating procedures (SOPs). flight path/glideslope angle airspeed landing configuration predicted touch down point Predict likelihood of an unstable approach based on extent of deviation of present and predicted approach profile from stable approach profile defined in SOP. Alerts If likelihood of unstable approach within next 20 seconds is high and persists for more than 5 seconds, generate alerts. C6. Wheels up landing On approach, monitor whether the An aircraft contacts the intended landing area aircraft has timely and correctly lowered with the landing gear retracted. A wheels-up landing gear. landing relates specifically to flight crew Predict likelihood of a wheels up landing an aircraft with the landing gear in a landing based on landing gear status and retracted state. This could be intentional due to distance between aircraft and airdrome. a mechanical issue or unintentional as the result Alerts of a distraction. If likelihood of wheels up landing within next 20 seconds is high and persists for more than 5 seconds, generate alerts. C7. Landing gear/Indication After landing, monitor landing gear When the landing gear or its component parts collapse, tyre damage/deflation, overheated or (including indications), has failed or exhibited smoking brakes, faults with floats and damage. Landing gear occurrences include: emergency flotation devices. after landing, landing gear collapse due Alerts to mechanical malfunction If likelihood of landing gear/indication landing gear indication problems occurrence immediately is high and persists for use of emergency gear extension more than 5 seconds, generate alerts. tyre damage/deflation overheated or smoking brakes faults with floats and emergency flotation devices C8. Incorrect configuration During take-off and landing phases, An aircraft system is incorrectly set for the monitor landing gear, flaps or slats current and/or intended phase of flight (take-off configuration, carburettor heat (if applicable), and landing phases). reverse thrust (if applicable). An aircraft system should not incorrectly set for Predict likelihood of incorrect the current and/or intended phase of flight configuration based on detected aircraft (take-off and landing phases). features. Incorrect configuration includes occurrences Alerts where flight crew: If likelihood of incorrect configuration within fail to extend the landing gear before next 20 seconds is high and persists for more landing (retract for amphibious operations) than 5 seconds, generate alerts. inadvertently retract the landing gear after landing incorrectly configure the flaps or slats incorrectly applies carburettor heat (if applicable) incorrectly applies reverse thrust (if applicable) C9. Ground proximity alerts/warnings During take-off and landing phases, A Ground Proximity warning or alert. monitor aircraft altitude above ground. An aircraft system should avoid a Ground Predict likelihood of ground proximity Proximity warning or alert. alerts/warnings based on comparison between aircraft altitude above ground and desired value. Alerts If likelihood of ground proximity alerts/ warnings within next 20 seconds is high and persists for more than 5 seconds, generate alerts. C10. Flight below Lowest Safe Altitude Similar to A3, A5, C4, C5, C9 (LSALT) An aircraft is operated below the designated or planned Lowest Safe Altitude (LSALT) for the in-flight conditions and phase of flight. Any occurrence that relates to an aircraft operating below the lowest safe altitude for the planned route, or area, in conditions other than day VMC. This includes: crew error to descend below the LSALT in IMC aircraft operating below LSALT without knowledge of terrain in the vicinity ATC instruction to descend or operate below the are LSALT or Radar LSALT aircraft that continue the approach below minimas with no visual reference to the runway. C11. Lost/unsure of position When flight crew are uncertain of the aircraft's position and/or request assistance from an external source. Occurrences where an aircraft requests navigational assistance from ATC or other external means (such as pilots of other aircraft), in determining their current position. C11. Collision with terrain (near and at Monitor aircraft tracks, and its distance airport) to other aircraft/terrain/ground Any collision between an airborne aircraft and vehicle/person/object. the ground, water or an object, where the flight Monitor awareness of one aircraft to the crew were aware of the terrain prior to the other aircraft/terrain/ground collision. vehicle/person/object, avoiding plan and/or C12. Collision/Near collision (near and at action. runway) Monitor damage to the fuselage, wings, An aircraft collides with another aircraft either or empennage. airborne or on the runway strip, or a vehicle or Predict likelihood of collision by person on the runway strip. monitoring distance between aircraft and other An aircraft comes into such close proximity aircraft/terrain/ground vehicle/person/object and with another aircraft either airborne or on the present and predicted aircraft tracks. runway strip, or a vehicle or person on the Alerts runway strip, where immediate evasive action If likelihood of collision within next 20 was required or should have been taken. seconds is high and persists for more than 2 Collision with terrain includes: seconds, generate alerts. impact with terrain (not including wires) from which the aircraft flies away airborne collisions with fences collision with objects on ground during take-off and landing or within the confines of a flight strip Collisions include: mid-air collisions collisions on the runway between two aircraft, or with a vehicle/person on a runway strip Near collision, considerations should be given to: one or both aircraft took significant avoiding action, or would have if time had permitted one or both aircraft received an unexpected TCAS RA aircraft tracks awareness of one aircraft to the other C14 Controlled flight into terrain (CFIT) Similar to A3, A5, C4, C5, C9 When a serviceable aircraft, under flight crew control, is inadvertently flown into terrain, obstacles or water without either sufficient or timely awareness by the flight crew to prevent the collision -occurs when pilot is in control of the aircraft and aware of the impending collision. The pilot's inadequate awareness of the terrain may result from a number of operational circumstances, including operating in IMC, at night, distractions, inadequate lookout, incorrect route flown, and in some cases may be the result of operating outside the tolerances of an instrument approach. C15 Loss of separation Monitor the The failure to maintain a recognised separation distance/proximity/separation between aircraft standard (vertical, lateral or longitudinal) with another aircraft/vehicle or person near and between aircraft that are being provided with an at airport. ANSP separation service includes a loss of: Predict likelihood of loss of separation procedural or surveillance/radar based on distance/proximity between aircraft separation standards with another aircraft/vehicle or person near and prescribed runway or wake turbulence at airport. separation standards Alerts runway proximity occurrences relating If likelihood of loss of separation within next 20 to a departing aircraft with another aircraft, seconds is high and persists for more than 5 vehicle or person occupying the same runway seconds, generate alerts. simultaneously (also to be coded as a Runway Incursion and, where applicable, Near Collision) visual separation by a pilot or air traffic controller in controlled airspace, if visual reference is lost C16 Aircraft separation - issues Airspace - Aircraft separation occurrences not specifically covered elsewhere. Aircraft separation - Issues occurrences where separation is a concern but does not meet the definition of Loss of Separation or Near collision. C17. Loss of separation assurance (LOSA) Similar to A3, A5, C4, C5, C9, C15, Where separation has been maintained but has C16 not been planned, actioned or monitored appropriately. LOSA is an occurrence where separation existed but: potential conflict was not identified; or separation was not planned or was inappropriately planned; or separation plan was not executed or was inappropriately executed; or separation was not monitored or was inappropriately monitored C18 Airspace infringement Similar to A3, A5, C4, C5, C9, C15, Where there is an unauthorised entry of an C16 aircraft into airspace for which a clearance is required. All occurrences, in which an aircraft enters controlled, restricted or a prohibited airspace without prior approval from the airspace owner is to be recorded as an Airspace Infringement. This includes incidents where an aircraft takes off from a designated position inside a controlled or restricted area before receiving approval to do so. This equally applies to aircraft departing from a controlled environment. C19 Operational non-compliance Non- compliance of an ANSP verbal or published instruction. These occurrence type relate specifically to flight crews not adhering to instructions issued by an ANSP, be it a verbal instruction or a clearance that relates to a published instruction. These instructions can relate to: aircraft heading route altitude busts flying the wrong SID or STAR or flying it incorrectly. C20 Airspace - other Airspace occurrences not specifically covered elsewhere. If an occurrence is coded as Airspace - Other, then a brief description of the actual event is recorded. D1. Animal strike Detect and monitor animals near and at AA collision between an aircraft and an animal. airport. Animal strikes occurrences include situations in Monitor distance/proximity between which the aircraft physically strikes any aircraft and animal(s) near and at airport. flightless animal. A near animal strike or a Predict likelihood of an animal strike suspected animal strike (where the pilot reports based on distance/proximity between aircraft that they may have hit an animal but no and animal(s) near and at airport. evidence is found) are also coded as animal Alerts strikes but as Events'. If likelihood of animal strike within next 20 seconds is high and persists for more than 5 seconds, generate alerts. D2. Bird strike Similar to D1 A collision between an aircraft and a bird. Bird strike occurrences include situations in which the aircraft is in flight, or taking off or landing. Bird strike also include occurrences where a bird carcass is found on a runway. A rejected take-off or go-around may be used as a preventative means of avoiding a bird. D3. Wildlife - Other Similar to D1 Wildlife related occurrences not specifically covered elsewhere includes: flying through insect plagues insects in pitot tubes etc reports of animals/birds on the aerodrome snakes on planes D4. Near encounter with remotely piloted Similar to A5, A6 and B3 aircraft D5 Turbulence/windshear/microburst Similar to A3, A5, C4, C5, C9, C15, Aircraft performance and/or characteristics are C16, C18 affected by turbulence, windshear or a microburst. When coding Turbulence/Windshear/ Microburst, the effect on aircraft performance or control must be clearly quantifiable, based on information such as: significant airspeed fluctuation significant altitude, or profile deviations significant changes in rate of climb or descent severity of encounter the degree of influence on aircraft control the degree to which the integrity of the airframe is affected if injury to occupants has occurred. D6 Unforecast weather Operations affected by weather conditions that were not forecast or not considered by the flight crew. Any aircraft operation that is affected by an unforecast weather phenomenon. Also includes weather conditions not considered prior to flight or during the flight by flight crew. Includes diversions, holding, missed approaches, or flight continues through adverse weather or visibility conditions. D7 Emergency evacuation When crew and/or passengers vacate an aircraft in situations other than normal and usually under the direction of the operational crew. An Emergency Evacuation is coded when there is a level of urgency to have all crew and passengers disembark as the result of an occurrence that places them at risk to serious injury or death. This can be achieved by any number of means, including: emergency slides, integrated aircraft stairs, aerobridge, or external stairs. D8. Fire Detect and monitor fire, fume, smoke, Any fire that has been detected and confirmed icing and lightning near and at airport, including in relation to an aircraft operation. aircraft. D9. Fumes Monitor distance between aircraft and When abnormal fumes or smells are reported on detected fire, fume, smoke. board the aircraft includes reports of abnormal Predict effects of fire, fume, smoke, smells not associated with normal aircraft icing, or lightning near and at airport. operations. Alerts D10. Smoke If fire, fume, smoke, icing or lightning When smoke is reported to be emanating from: is adversely affecting operating aircraft near and a) inside the aircraft; or at airport, generate alerts. b) an external component of the aircraft; or If fire, fume, smoke, icing or lightning c) a smoke alarm activates. is detected, inform relevant stakeholder, airport Smoke occurrences relate specifically to fire service, to inspect/extinguish. non-normal situations whereby crew, ground staff or passengers detect smoke that is not associated with the normal operation of the aircraft. D11. Icing Accumulation of ice on aircraft that adversely affects aircraft controllability D12. Lightning Strike The aircraft, or another object, is struck by lightning E1 Radar/surveillance Similar to A3, A5, C4, C5, C9, C15, Any faults or deficiencies in the operation of a C16, C18 radar or surveillance system used for the purpose of separating aircraft in the air or on the ground. This occurrence type relates specifically to failed radar or surveillance services, including ADS-B ground stations, where no redundancy exists, and ANSP services revert to a procedural environment. Where available, the record should indicate the length of time the facility was out of service. Where redundancy is available and there has been little or no effect on operations then the occurrence is to be classified as an Event E1. Runway lighting Detect and monitor runway lighting. Any faults or deficiencies associated with the Predict effects of runway lighting operation of runway lighting. This occurrence faults/deficiencies on operating aircraft near and type covers all types of runway lighting issues at airport. necessary for the safe operation of aircraft Alerts during the take-off and landing phases of flight. If runway lighting faults/deficiencies is This includes: adversely affecting operating aircraft near and approach and slope guidance lighting at airport, generate alerts. (PAPI & HIRL) If runway lighting faults/deficiencies is runway edge and centre lighting detected, inform relevant stakeholder, airport Pilot Activated Lighting (PAL) where the fault engineering team, to inspect/repair. is linked to runway ground equipment E2. Infrastructure - other Similar to B1, B2, B3, B5, B6, B16 Infrastructure related occurrences not specifically covered elsewhere. If an occurrence is coded as Infrastructure - Other, then a brief description of the actual event is recorded in the accompanying text box.

    [0136] Table 3 sets out an example of the occurrence types and groups that occur in an aviation environment particularly near and at airport (left column), such as runway (A1 to A7), ground operations (B1 to B14), aircraft control (C1 to C20), environment (D1 to D12), infrastructure (E1 to E3) occurrence groups. Multiple occurrence types can be monitored within each occurrence group category. In one example, the occurrence type runway excursion A1 is one of the occurrence types that are classified under runway occurrence group. These occurrence types are level 3 occurrence types, which are defined and used by Australian Transport Safety Bureau (ATSB). Advantageously, the system may be configured to monitor up to 59 ATSB level 3 occurrence types, i.e. A1 to E3 as exemplified in Table 3, in comparison with the five occurrence types which are typically monitored using current aviation safety monitoring systems.

    [0137] In the right column of Table 3, there are shown detection and tracking multiple objects data processing capability and brief safe operation criteria that are required for each occurrence group, including object types, classes, different physical (both current and predicted) properties of each monitored object, and the types of risks and accompanying safe operation criteria that is associated with each occurrence type. Table 4 provides additional details into the particular safe and unsafe operation criteria (left column) for each of the occurrence types and in the right hand column there is provided the examples of assessment criteria/method for each of the safety operation criteria.

    [0138] For example, in the example of the occurrence type A1 runway excursion illustrated in FIGS. 3 to 6, the system 2 and method 200 is first configured to receive sensor information from the at least two sensors 26, 28, 30, i.e. the LiDAR 26 and camera sensors 28, from at least one monitoring unit 22, located in at least one location in the aviation environment in step 202. The system 2, the processing system 4 in particular, is configured to fuse the two types of sensors' information with temporal and spatial information

    [0139] Further in step 204 the system 2 is configured to process the sensor information including using the fused information to identify/classify/detect at least one object, such as the aircraft 16 and runway 40. Further the system 4 can calculate the at least one objects' physical properties, and to predict the at least one objects' physical properties. For example, the aircraft's position, travel direction, velocity, acceleration, altitude and attitude is monitored as well as the distance between aircraft 16 of interest and object of interest, e.g. the runway 40 in particular its' surface, boundary 49, markings 50, centreline 52, runway threshold 48 and to calculate runway distance remaining, distance between aircraft and runway boundaries, centreline and the like.

    [0140] Additional information 54 can be received by the processing system 4 to assist and/or facilitate calculation of the objects' physical properties and estimation/prediction of their physical properties and/or safe operation criteria, including runway data, such as length, boundaries, entries and exits, surface characteristics such as material or friction coefficients, and/or surface conditions such as wet, ice/snow, metrological data such as wind, temperature and the like, and aircraft data, such air craft type and capabilities/characteristics, weight, flying phase and/or intended or reference position or motion.

    [0141] Further in a next step 206 of the method 200, the system 2 is configured to measure or calculate an estimate or prediction of the particular physical properties of the aircraft 16 and runway 40 which may relate to a particular predetermined safety criteria, i.e. A1. For aircraft landings, as illustrated in FIG. 4, the system 2 is configured to monitor the aircraft approach flight path from when the aircraft 16 is 50 metres above the ground 31, to measure and/or calculate an estimate or prediction of the touch-down point 32, including by measuring and calculating predicted location, travel direction, velocity, deceleration and altitude as exemplary physical properties. After the aircraft 16 has touched down on the runway 40, the system 2 is configured to track aircraft position 33 along a tracked path 37, to determine the predicted path 38, based on the measured and predicted aircraft position and speed, and to calculate and predict where the aircraft's speed will become low enough to ensure a safe stop to a safe stopping position 34 before the end of the runway 40 and expected run-way exit point 35.

    [0142] As illustrated in FIG. 5, for aircraft take-off, the system 2 is configured to monitor and/or track current aircraft location 63 of the aircraft 16, and/or measure and/or calculate an estimate or prediction of the lift-off position 62, last safe stopping point 64, after it has started from its take-off roll position 61, and before it commences its airborne flight path 65.

    [0143] In the example shown in FIG. 6, to monitor for the risk of runway veer-off, the system 2 is configured to monitor and/or track current aircraft location 73 of the aircraft 16, and/or measure and/or calculate an estimate or prediction of the position of the aircraft on the runway (predicted path) 78, aircraft position deviation from runway centreline 52, distance between aircraft and runway boundary 49, and predicted position where risk of veer-off is high 72, and predicted veer-off position 74.

    [0144] The system 2 is also configured to store the particular safe operation criteria in step 208 such as the calculated safe lift-off position for a particular aircraft type, for example, under specific aircraft loading, runway and meteorological conditions during take-off, calculated safe stopping position i.e. where the aircraft speed becomes low enough to ensure a safe stop before the end of the runway and/or aircraft can safely exit from the runway for a particular aircraft type under specific aircraft loading, runway and meteorological conditions during landing. The system 2 can also be configured to calculate the acceptable limits for the lift-off, veer off, touch-down and safe stopping positions, i.e. acceptable runway distance remaining, and/or to calculate and predict the safe operation criteria as required.

    [0145] The system 2, is then configured to compare the measured or predicted physical properties of the aircraft 16 and runway 40 to the safe operation criteria to determine the potential runway excursion risks. In particular the system 2 in step 212 can predict the likelihood of runway excursion by monitoring distance between aircraft landing gears/fuselage/wingtip and runway side boundary for veer off and by monitoring runway distance remaining for runway overrun. If the comparison shows that the measured and predicted physical properties of the aircraft and runway are acting within safe operating criteria, then the system 2 can determine that the likelihood of risk of runway excursion is low and an indication/alert may be generated to a user to confirm safe aviation operation.

    [0146] Alternatively, the system 2 is configured to determine that the comparison shows that risk of runway excursion is medium or high, i.e. runway excursion may occur in the next 15 seconds, or in the next 5 seconds, and the system is further configured to transmit at least one alert to at least one user accordingly. The user(s) could include aviation traffic control (ATC), pilots, emergency response team, and the like. Finally, if an excursion has occurred, the system 2 is configured to send an alert for at least one user, particularly emergency response teams and relevant authorities.

    [0147] Lastly, as illustrated in FIG. 3, the system 2 in step 214 is configured to suggest corrective or mitigation actions if necessary, i.e. if it has been determined that the risk of runway excursion is not low but is medium or high to at least one appropriate user. For example, for take-off, the system 2 is configured to send an alert to at least one pilot of the aircraft to adjust power settings to accelerate the take-off or to abort the take-off. Accordingly, for landing, the system 2 can send an alert to at least one pilot to adjust power settings to slow down (e.g. reverse thrust), deploy spoiler and/or increase tyre braking, or conduct go around or touch and go. Similarly, for a runway veer-off, the alert could be sent to at least one pilot to steer the aircraft back to centreline from an off-centreline location. Alternatively, the system 2 can recommend deployment of an engineered materials arresting system or an aircraft arresting system, i.e. a net-type aircraft barrier or an alternative system or apparatus having an equivalent function, to prevent runway overrun.

    [0148] The system 2 in step 216 is also able to receive information from existing safety nets to facilitate processing of information and calculation of measured operational physical properties and prediction thereof and to act as a redundancy. For example, the system can be configured to receive information from runway overrun protection systems (ROPSI).

    [0149] In a further example of the system 2 and method 300 as discussed according to preferred embodiments of the present invention, as illustrated in FIGS. 7 and 8, the risks associated with ground operations (B1 to B16) in the aviation environment near and at the airport can be more particularly monitored including taxiing collision/near collision, foreign object damage/debris 55, objects falling from aircraft 56, jet blast/propeller/rotor wash 57, fire/fume/smoke 58, fuel leaks 59, damage to aircraft fuselage/wings/empennage 60 and the like however in this example taxiing collision/near collision B3 is discussed in more detail.

    [0150] For example, in the more specific example of the occurrence type B3 taxiing collision/near collision illustrated in FIGS. 7 and 8, the system 2 is first configured in step 302 to produce, transmit and/or receive sensor information from the at least two sensors, i.e. the LiDAR and camera sensors 26, 28, in the form of multiple monitoring units 22 from multiple locations in the aviation environment. The system 2, the processing system 4 in particular, is configured to fuse the two sensors' information from each monitoring unit 22 by apply a time-syncing process and/or a sensor calibration process.

    [0151] Further the system 2 in step 304 is configured to use the fused information to detect and identify at least one object, such as the aircraft(s) 16, ground vehicles and crew 18, 20, and airport infrastructure such as the boarding gates and the like, to calculate the at least one objects' physical properties, and to predict the at least one objects' physical properties. For example, the aircrafts' position, travel direction, velocity, acceleration, altitude and attitude is monitored as well as the distance between aircraft of interest and object of interest, e.g. the ground vehicles and crew, boarding gates, gate boundaries and the like.

    [0152] Additional information 54 can be received by the processing system 4 to assist and/or facilitate calculation of the objects' physical properties and estimation/prediction of their physical properties and/or safety operation criteria, including runway data, such as boarding gates and bridges, apron and ramp area boundaries, surface characteristics such as material or friction coefficients, and/or surface conditions such as wet, ice/snow, metrological data such as wind, temperature and the like, and aircraft data, such air craft type and capabilities/characteristics, weight, flying phase and/or intended or reference position or motion, and ground crew/vehicle data.

    [0153] Further in a next step 306 of the method 300, the system 2 is configured to measure or calculate an estimate or prediction of the physical properties of the aircraft(s) 16, airport infrastructure and ground vehicles/crew 18, 20. For an aircraft taxiing to and from the boarding gates and bridges, as illustrated in FIG. 8, the system 2 is configured to monitor, measure and/or calculate an estimate or prediction of the path(s) of aircraft(s) 16 taxiing to and from the boarding gates, and the movement of nearby ground crew and vehicles. The monitored, measured and/or calculate physical properties include position, speed, travel direction, track and acceleration.

    [0154] As illustrated in FIG. 8, the system 2 is configured to measure and/or calculate an estimate or prediction of the distance between the aircraft(s) and any ground crew/vehicles and airport infrastructure to monitor any risk of collisions or near-collisions therebetween.

    [0155] The system 2 in step 308 is also configured to store the particular safe operation criteria such as the defined and/or calculated safe distances between the objects, i.e. aircraft 16, ground vehicles/crew infrastructure 18, 20. The system 2 can also be configured to calculate the acceptable limits for the same and/or to calculate and predict the safe operation criteria as required.

    [0156] The system 2, in the next step 312, is then configured to compare the measured or predicted physical properties of the aircraft 16 and other objects to the safe operation criteria to determine the potential collision risks. In particular the system 2 can predict the likelihood of collisions or near collisions by monitoring distance between any two or more objects, i.e. the distance between the aircraft(s) 16 and any ground crew/vehicles 18, 20 and airport infrastructure 42, 44. If the comparison shows that the measured and predicted physical properties of the aircraft 16, ground crew/vehicles 18, 20 and airport infrastructure are acting within safe operating criteria, then the system 2 in step 312 can determine that the likelihood of risk of runway excursion is low and an indication/alert may be generated to at least one user to confirm safe aviation operation.

    [0157] Alternatively, the system 2 is configured to determine that the comparison shows that risk of collision or near collision is medium or high, i.e. runway excursion may occur in the next 120 seconds, or in the next 20 seconds, and the system is further configured to transmit at least one alert to at least one user accordingly i.e. yellow alert or red alert. The user(s) could include aviation traffic control (ATC), pilots, emergency response team, and the like. Finally, if a collision has occurred, the system is configured to send an alert for at least one user, particularly emergency response teams and relevant authorities.

    [0158] Lastly, as illustrated in FIG. 7, the system 2 in step 314 is configured to suggest corrective or mitigation actions if necessary, i.e. if it has been determined that the risk of runway collision or near collision is not low but is medium or high to an appropriate user. For example, the system 2 is configured to send an alert to a pilot, ground personnel 20 to slow/stop and/or conduct collision avoidance measures.

    [0159] The system 2 in step 316 is also able to receive information from existing safety nets i.e. traffic conflicts by STCA to facilitate processing of information and calculation of measured operational physical properties and prediction thereof and to act as a redundancy.

    [0160] In another further example of the system 2 and method 400 according to preferred embodiments of the present invention, FIGS. 9 and 10 illustrate the aircraft and runway control (A3, A5, C5, C6). More particularly using this method 400 the system 2 can be used to monitor risks associated with the aviation activities in the aviation environment near and at the runway 40 including runway undershoots, depart/approach/land wrong runway, unstable approach and wheels up landing however in this example wheels up landing and unstable approach is discussed in more detail below. FIG. 10 illustrates the application of the system 2 which tracks the path 87 of the aircraft 16, monitors the current location 83 of the aircraft, and predicts the acceptable spatial limits for stable approach 81 as well as a predicted approach flight path 88 and touch down point 82.

    [0161] The system 2 is first configured in step 402 to receive sensor information from the at least two sensors, i.e. the LiDAR and camera sensors 26, 28, in the form of multiple monitoring units 22, in the aviation environment. The system 2, the processing system 4 in particular, is configured to fuse the two sensors' information from each monitoring unit 22 by a time-syncing process and/or sensor calibration process.

    [0162] Further the system 2 in step 404 is configured to use the fused information to detect and/or identify at least one object, such as the aircraft 16, to detect and classify at least one object feature, such as the aircraft landing gear status, i.e. landing gear 53, in an extended or a retracted position, to calculate the at least one objects' physical properties, and to predict the at least one objects' physical properties. For example, the aircraft's position, travel direction, velocity, acceleration, altitude and attitude are monitored.

    [0163] Additional information 54 can be received by the processing system 4 to assist and/or facilitate calculation of the objects' physical properties and estimation/prediction of their physical properties and/or safety operation criteria, including runway data, such as length, boundaries, entries and exits, surface characteristics such as material or friction coefficients, and/or surface conditions such as wet, ice/snow, metrological data such as wind, temperature and the like, and aircraft data, such air craft type and capabilities/characteristics, weight, flying phase and/or intended or reference position or motion.

    [0164] Further in a next step 406 of the method 400, the system 2 is configured to measure or calculate an estimate or prediction of the physical properties of the aircraft 16 and runway 40, in the system 2 is configured to measure and/or calculate an estimate or prediction of the approach flight path, tracked current aircraft location, deviation of path profile parameters such as lateral and vertical profile, airspeed, bank angle, altitude, vertical speed, altitude and attitude.

    [0165] For wheels up landing, as illustrated in FIGS. 9 and 10, the system is configured to particularly detect/classify/measure the configuration of the landing gear 53, such as whether the landing gear is extended/deployed partially or fully, extended/deployed in a timely way or is still in a retracted position.

    [0166] The system 2 in step 408 is also configured to store the particular safe operation criteria such as for wheels up landing, whether on approach, the spatial position along the approach flight path of the aircraft 16 at which the landing gear should be fully extended/deployed to achieve safe touchdown/landing. For unstable approach, acceptable deviation of measured flight path from the intended/authorised/ideal flight path. The system 2 can also be configured to calculate the acceptable limits thereof and/or to calculate and predict the safe operation criteria as required.

    [0167] The system 2 in the next step 412, is then configured to compare the measured or predicted physical properties of the aircraft and runway to the safe operation criteria to determine the potential risks. In particular the system can predict the likelihood of incorrect aircraft landing configuration by monitoring the landing gear configuration. The system can also predict the likelihood of unstable approach by monitoring the approach flight path. If the comparison shows that the measured and predicted physical properties of the aircraft and landing gear configuration are acting within/complying safe operating criteria, then the system can determine that the likelihood of risk of wheels up landing and/or unstable approach is low and an indication/alert may be generated to at least one user to confirm safe aviation operation.

    [0168] Alternatively, the system 2 is configured to determine that the comparison shows that risk of wheels up landing and/or unstable approach is medium or high, i.e. wheels up landing and/or unstable approach may occur in the next 120 seconds, or in the next 20 seconds, and the system is further configured to transmit at least one alert to at least one user accordingly i.e. yellow alert or red alert. The user(s) could include aviation traffic control (ATC), pilots, emergency response team, and the like. Finally, if a wheels up landing has occurred, the system is configured to send an alert for at least one user, particularly emergency response teams and relevant authorities.

    [0169] Lastly, as illustrated in FIG. 9 in step 414, the system 2 is configured to suggest corrective or mitigation actions if necessary, i.e. if it has been determined that the risk of wheels up landing and/or unstable approach is not low but is medium or high to an appropriate user. For example, for wheels up landing, the system is configured to send an alert to at least one user, e.g. pilot, to check and/or advise landing gear configuration. For unstable approach, the system is configured to send an alert to at least one user, e.g. pilot, to check and/or advise the measured deviation from ideal flight path, and the aircraft to conduct a go around or touch and go.

    [0170] The system 2 in step 416 is also able to receive information from existing safety nets to facilitate processing of information and calculation of measured operational physical properties and prediction thereof and to act as a redundancy. For example, the system 2 can be configured to receive information from High Energy Approach Monitoring Systems (ROPSI).

    [0171] The system can provide real-time monitoring of aviation activities, detection of unsafe aviation activity and generation of alerts, which can be displayed on at least one standalone screen or can be integrated with existing systems located in at least a cockpit of said aircraft, air traffic control towers/centres, ground control locations and airport emergency response team locations. The display format may include 3-D map and panoramic view.

    [0172] The system and methods described above provide one or more of the following advantages including improvement in aviation safety, operation efficiency, capacity, operating cost efficiency, environment and security. Specifically, the advantages include the following: enhanced situation awareness of unsafe aviation activities to human operators and operating systems, e.g. Air Traffic Control officers, pilots, aircraft on board systems that control the aircraft and emergency response team: awareness of all objects and activities within the aviation operating environment near and at airport; prompt detection and awareness (within seconds) of deviation from and/or violation of safe aviation operation criteria; human operators and/or operating systems can immediately assess the detected and identified unsafe aviation activities, and implement appropriate corrective actions; prevention of aviation safety occurrences or reduction of severity/cost of aviation safety occurrences; increased redundancy to the existing technologies and procedures that detect/identify/prevent/mitigate unsafe aviation activities; a more cost-effective solution/technique/system compared to existing systems/technologies/solutions; reduced reliance on human involvement, e.g. human observation at Air Traffic Control; minimum changes to current procedures or workload.

    INDUSTRIAL APPLICABILITY

    [0173] It is apparent from the above, that the arrangements described are applicable to aviation industries, and related industries, and the processes, systems and equipment therefor.

    GENERAL STATEMENTS

    Embodiments

    [0174] Reference throughout this specification to one embodiment or an embodiment means that a particular feature, structure or characteristic described in connection with the embodiment is included in at least one embodiment of the present invention. Thus, appearances of the phrases in one embodiment or in an embodiment in various places throughout this specification are not necessarily all referring to the same embodiment, but may. Furthermore, the particular features, structures or characteristics may be combined in any suitable manner, as would be apparent to one of ordinary skill in the art from this disclosure, in one or more embodiments.

    [0175] Similarly it should be appreciated that in the above description of example embodiments of the invention, various features of the invention are sometimes grouped together in a single embodiment, figure, or description thereof for the purpose of streamlining the disclosure and aiding in the understanding of one or more of the various inventive aspects. This method of disclosure, however, is not to be interpreted as reflecting an intention that the claimed invention requires more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive aspects lie in less than all features of a single foregoing disclosed embodiment. Thus, the claims following the Detailed Description are hereby expressly incorporated into this Detailed Description, with each claim standing on its own as a separate embodiment of this invention.

    [0176] As used herein in the specification and claims, including as used in the examples and unless otherwise expressly specified, all numbers may be read as if prefaced by the word about or approximately, even if the term does not expressly appear. The phrase about or approximately may be used when describing magnitude and/or position to indicate that the value and/or position described is within a reasonable expected range of values and/or positions. For example, a numeric value may have a value that is +/0.1% of the stated value (or range of values), +/1% of the stated value (or range of values), +/2% of the stated value (or range of values), +/5% of the stated value (or range of values), +/10% of the stated value (or range of values), etc. Any numerical values given herein should also be understood to include about or approximately that value, unless the context indicates otherwise. For example, if the value 10 is disclosed, then about 10 is also disclosed. Any numerical range recited herein is intended to include all sub-ranges subsumed therein. It is also understood that when a value is disclosed that less than or equal to the value, greater than or equal to the value and possible ranges between values are also disclosed, as appropriately understood by the skilled artisan. It is also understood that each unit between two particular units are also disclosed. For example, if 10 and 15 are disclosed, then 11, 12, 13, and 14 are also disclosed.

    Different Instances of Objects

    [0177] As used herein, unless otherwise specified the use of the ordinal adjectives first, second, third, etc., to describe a common object, merely indicate that different instances of like objects are being referred to, and are not intended to imply that the objects so described must be in a given sequence, either temporally, spatially, in ranking, or in any other manner.

    Specific Details

    [0178] In the description provided herein, numerous specific details are set forth. However, it is understood that embodiments of the invention may be practiced without these specific details. In other instances, well-known methods, structures and techniques have not been shown in detail in order not to obscure an understanding of this description.

    Terminology

    [0179] The terms in the claims have the broadest scope of meaning they would have been given by a person of ordinary skill in the art as of the relevant date.

    [0180] The term associate, and its derivatives (e.g. associating) in relation the combination of data includes the correlation, combination or similar linking of data.

    [0181] The term data fusion, fusing and like terms are intended to refer to a multi-level process dealing with the association, correlation, combination of data and information from single and multiple sources to achieve refined position, identify estimates and complete and timely assessments of situations, risks and their significance.

    [0182] The terms a and an mean one or more, unless expressly specified otherwise

    [0183] Neither the title nor any abstract of the present application should be taken as limiting in any way the scope of the claimed invention.

    [0184] Where the preamble of a claim recites a purpose, benefit or possible use of the claimed invention, it does not limit the claimed invention to having only that purpose, benefit or possible use.

    [0185] In the present specification, terms such as part, component, means, section, or segment may refer to singular or plural items and are terms intended to refer to a set of properties, functions or characteristics performed by one or more items having one or more parts. It is envisaged that where a part, component, means, section, segment, or similar term is described as consisting of a single item, then a functionally equivalent object consisting of multiple items is considered to fall within the scope of the term; and similarly, where a part, component, means, section, segment, or similar term is described as consisting of multiple items, a functionally equivalent object consisting of a single item is considered to fall within the scope of the term. The intended interpretation of such terms described in this paragraph should apply unless the contrary is expressly stated or the context requires otherwise.

    [0186] The term connected or a similar term, should not be interpreted as being limitative to direct connections only. Thus, the scope of the expression an item A connected to an item B should not be limited to items or systems wherein an output of item A is directly connected to an input of item B. It means that there exists a path between an output of A and an input of B which may be a path including other items or means. Connected, or a similar term, may mean that two or more elements are either in direct physical or electrical contact, or that two or more elements are not in direct contact with each other yet still co-operate or interact with each other.

    Comprising and Including

    [0187] In the claims which follow and in the preceding description of the invention, except where the context requires otherwise due to express language or necessary implication, the word comprise or variations such as comprises or comprising are used in an inclusive sense, i.e. to specify the presence of the stated features but not to preclude the presence or addition of further features in various embodiments of the invention.

    [0188] Any one of the terms: including or which includes or that includes as used herein is also an open term that also means including at least the elements/features that follow the term, but not excluding others. Thus, including is synonymous with and means comprising.

    SCOPE OF INVENTION

    [0189] Thus, while there has been described what are believed to be the preferred embodiments of the invention, those skilled in the art will recognize that other and further modifications may be made thereto without departing from the spirit of the invention.

    [0190] Functionality may be added or deleted from the block diagrams/flow charts, and operations may be interchanged among functional blocks. Steps may be added or deleted to methods describe within the scope of the present invention.

    [0191] Although the invention has been described with reference to specific examples, it will be appreciated by those skilled in the art that the invention may be embodied in many other forms.