Commercial Farm Optimization Utilizing Simulation, Remote Sensing, and Farmer Inputs

20220124960 · 2022-04-28

    Inventors

    Cpc classification

    International classification

    Abstract

    Briefly, an advanced data collection and processing system is provided to collect multiple types of data from farm terrain to drive a farm management processes, including a crop yield simulation tool. The system has disparate sensors mounted on a vehicle, such as a ground vehicle or airplane, which collects data from GPS, RADAR, camera, thermal, LiDAR and spectral scanners for an area of interest on a farm terrain. The system also collects data from public sources and the farm manager, which enable the simulation tool to accurately predict crop growth and maturity.

    Claims

    1. A data collection system for collecting and analyzing farm terrain data, comprising: location data for a plurality of measured spots that aggregate to represent farm terrain; ground penetrating RADAR data for the measured spots of farm terrain; microclimate data for the measured spots of farm terrain; soil moisture and soil type data for the measured spots of farm terrain; environmental data for the farm terrain; crop type data for the farm terrain; and wherein the data collection tool provides the location data, radar data, microclimate data, soil moisture data, soil type data and environmental data to simulate growth and maturity of the crop type for the farm terrain.

    2. The data collection system according to claim 1, further comprising LiDAR data, which is further provided to the farm crop simulation tool.

    3. The data collection system according to claim 1, further comprising data from a plurality of RADARs, which is further provided to the farm crop simulation tool.

    4. The data collection system according to claim 1, wherein the RADAR deploys multiple antennas in order to electronically steer the measurement location.

    5. The data collection system according to claim 1, wherein the RADAR is a synthetic aperture RADAR which is controlled by the controller to electronically steer the measurement location

    6. A vehicle for collecting data for a plurality of measured volumes that aggregate to represent farm terrain, comprising: instruments for collecting information regarding each measured volume, comprising: a GPS receiver for collecting location data; ground penetrating RADAR for collecting soil type data and soil moisture data; optical camera for collecting visible information; thermal imager; and spectrographic imager; a controller connected to the instruments; and wherein the controller, according the each instrument's latency and resolution, determines when each instrument will be triggered to collect information regarding each measured volume.

    7. The vehicle according to claim 6, wherein the instruments further comprise LiDAR, and the controller further triggers the LiDAR to collect data according to its latency and resolution.

    8. The vehicle according to claim 6, wherein the instruments further comprise multiple RADAR antennas for signal processing to create a larger aperture, and the controller further triggers each RADAR to collect data according to its latency and resolution.

    9. The vehicle according to claim 6, wherein the instruments further comprise multiple RADAR antennas for signal processing to synthetic aperture, and the controller further triggers each RADAR to collect data according to its latency and resolution.

    10. The vehicle according to claim 6, wherein the vehicle is a ground vehicle.

    11. The vehicle according to claim 6, wherein the vehicle is an airplane, drone or balloon.

    12. The vehicle according to claim 11, further comprising stabilization system for optical instruments, further comprising: passive stabilization feet connected to the vehicle; a stabilization table attached to the stabilization feet; and a plurality of optical instruments mounted on the stabilization table.

    13. The vehicle according to claim 12, wherein the plurality of optical instruments are selected from the group consisting of: optical camera, spectrometers, LiDAR and thermal sensor.

    14. A method of correlating disparate sensors for a data collection system that is collecting farm terrain data for an area of interest, comprising: connecting a plurality of sensing instruments connected to a common controller, the instruments further comprising two or more of: ground penetrating RADAR; optical camera; thermal imager; and a spectrographic imager; determining, for each instrument, the number of collection spots needed to cover the area of interest; triggering, using the controller, each instrument according to its latency and spot size; collecting data from each instrument, and time stamping the collected data; comparing the data collected from the optical camera to a stored reference image to generate an error vector between the actual image data and the reference image data; and adjusting location data for the collected data using the error vector.

    15. The method according to claim 14, further including stamping the collected data with GPS location information.

    16. The method according to claim 14, further including adjusting the stamped GPS location data by the error vector.

    17. The method according to claim 14, further comprising changing the number of measurement spots for the RADAR according to the RADAR's height above the area of interest.

    18. The method according to claim 14, wherein the instruments further comprise LiDAR, and the controller further triggers the LiDAR to collect data according to its latency and resolution.

    19. The method according to claim 14, wherein the instruments further comprise multiple RADAR antennas for signal processing to create a larger aperture, and the controller further triggers each RADAR to collect data according to its latency and resolution.

    20. The method according to claim 14, wherein the instruments further comprise multiple RADAR antennas for signal processing to synthetic aperture, and the controller further triggers each RADAR to collect data according to its latency and resolution.

    Description

    BRIEF DESCRIPTION OF THE DRAWINGS

    [0028] FIG. 100 is an illustration of farm practice, farm soil and farm crop yield simulation in accord with an embodiment of the present invention.

    [0029] FIG. 200 is an illustration of a farm croup management process flow diagram in accord with an embodiment of the present invention.

    [0030] FIG. 300 is a crop management information flow chart in accord with an embodiment of the present invention.

    [0031] FIG. 400 is an illustration of a cloud simulator framework and data flow in accord with an embodiment of the present invention.

    [0032] FIG. 500 is a diagram of the vehicle complement of equipment in accord with an embodiment of the present invention.

    [0033] FIG. 600 is a table of remote sensing equipment relationships in accord with an embodiment of the present invention.

    [0034] FIG. 700 is an illustration of a controller alignment of disparate sensors in accord with an embodiment of the present invention.

    [0035] FIG. 800 is an illustration of an optical instrument stabilization and housing design in accord with an embodiment of the present invention.

    [0036] FIG. 900 is an illustration of position vector error correction in accord with an embodiment of the present invention.

    DESCRIPTION

    [0037] Governments have spent billions in an attempt to support increased crop yields, crop modelling, and a wide variety of agriculture related activities. This spending is typically developed in universities and private institutions who utilize it for a wide variety of things. Yet to date, wide adoption of crop yield optimization tools are not used by even the large corporate farms in a standalone fashion and are rarely used by small farmers.

    [0038] Currently the largest such simulation tool is APSIM or Agricultural Production Systems sIMulator. APSIM Initiative is a collaboration between Australia's national science agency CSIRO, the Queensland Government, The University of Queensland, University of Southern Queensland, and internationally with AgResearch Ltd in New Zealand, and Iowa State University in the United States. From humble beginnings twenty years ago, APSIM is evolved to a framework containing many of the key models required for modelling crop growth on single and multi-field farms.

    [0039] With 20 years of development and worlds most used model, in 2017 APSIM adoption was <487 users in Australia and <250 users in the United States. Unfortunately, most commercial farms do not possess enough data to run APSIM nor do they know what to do with the data once it is calculated. This demonstrates the large gap between good intentioned governments and the fanner.

    [0040] The invention solves the problem by opening up a simple discussion with individual farm manager/agronomists first utilizing Farm Practice Optimization 110, then Farm Soil Optimization 120. These inputs are required to run the whole Farm Management Optimization module 130.

    [0041] Farm Soil Optimization module is comprised of four sub-modules and is designed to open a discussion with the farm manager to inform and optimize soil management, and irrigation, and soil nutrient management on his/her farm. This module uses inputs from the first module, Farm Practice optimization 110 and focuses on soil type 122, surface flatness 124, soil drainage 126, and nutrient application 128. The software will deliver maps of soil type and soil moisture as a function of time from previous farm scans held in the cloud database. The purpose of this module is to inform and help the farm manager optimize the farming process.

    [0042] Soil type 122 is the first sub module delivered in the farm soil optimization module. Images of the farm soil type in three dimensions are provided to give the farm manager detailed description of what is happening beneath the soil surface. Software will identify possible issues with clay or other impediments given prior inputs such as surface flatness, irrigation methodology etc.

    [0043] Soil Drainage 126 is the second sub module delivered in the farm soil optimization module. The module delivers review water transport (movement vs time) in the farmland. Highlighted are insufficiently irrigated regions in soil plots with suggestions to add either adjust irrigation or add additional drainage. In areas where fanners are adding tiles (drainage pipes) to drain water away from the farmland, the simulator will suggest where the old system may not be working and suggest areas which require additional drainage.

    [0044] Surface Flatness 128 is the third sub-module. This module delivers the farm surface in three dimensions and is very important for fanners who flood irrigate. This very accurate representation will suggest when areas of the farm are not level or have adequate slope.

    [0045] Nutrient application 129 is the last sub-module and prepares the farm for the use of the crop yield simulation. Nutrient application and soil preparation prior to planting is very important as it drives both production yield and over application of fertilizer becomes unwanted farm runoff.

    [0046] Optimization of crop yield 130 requires optimization of farm practices 110, and optimization of land use 120 prior to use. Information required of the fanner to complete these modules is easily delivered. Once completed, the whole farm simulator has enough information to operate thus delivering the ability to suggest optimum seeding dates based on seed type and manufacturer, or profit as a function of seed manufacturer to name two.

    [0047] Crop Simulation 130 is the third and last module delivered by the system as shown in FIG. 100, which shows a schematic of crop growth inside soil from soil prep through post-harvest. Simulator calculations 132 shows a schematic diagram of measurements made by the sensors. Maturity model 134 shows crop maturity vs time as simulated in the simulator. Inputs 136 shows key inputs form separate sources including farm manager inputs. Lastly yield 138 shows resultant farm yield. The resultant simulator allows the farmer the luxury of planning his crop using a plant date and models of climate and soil type and soil moisture as a function of time to simulate the growth of his crop through harvest. Variables such as irrigation times and amounts, irrigation methodology, nutrient application plan and rate, chemical application plan are all loaded into the simulator and adjust yield output thus allowing the farm manager the ability to optimize his yield and profit.

    [0048] Referring now to FIG. 200, a farm crop management process flow diagram is shown. Cloud 210 is a representation of the cloud or a simplified version of a computer with storage and access to multiple data sets as defined later in this document. Inputs 220 to the cloud computer 210 are updated weather information for the area under question. It will be appreciated that other inputs may be used. The sensor package in the scanner is deployed in an aircraft or ground vehicle and generates scan input 215 data for the farm terrain, and when it has completed its measurements data is uploaded into the cloud 210 for processing. Farmer input 239 is how the fanner informs the tool about his farm and farming processes. Once information is uploaded into the cloud 210 or computer, the computer can prepare advise for farm optimization 230. Farm optimization 230 is broken down into three areas, farm practice, land management, and crop management as depicted in the figure.

    [0049] Referring now to FIG. 300, a crop management information flow chart is illustrated. Farm inputs 310 summarizes fanner inputs to simulator. These inputs include farm practices, pollination methodology, farm application of: irrigation, nutrients such as fertilizer, and chemical such as pesticides. These inputs also include surface organic matter if any and farm location.

    [0050] Cloud data set 320 summarizes remote sensing cloud data set. This data set includes soil type, soil moisture, water applied due to rainfall, and dew, measured nutrient content, a model of microclimate and determination of infestation status.

    [0051] Outputs to fanner 330 summarizes the output of the cloud computer. These outputs come in two types; the user can interrogate some measured data (not shown in picture) The second types of outputs are outputs delivered from the simulation tool. These outputs include suggested irrigation quantity and date, nutrient application quantity and date, and lastly chemical infestation status, climate inversion dates, and suggested times for application.

    [0052] Referring now to FIG. 400, a data flow and cloud simulator process are diagramed. Individual raw scanned data is delivered to the cloud and stored in the data scan memory 410. Data is read in by the scan developer computer which selects individual scan data required to deliver a single output parameter using the algorithm defined in FIG. 600. A simple example of this is to look at the location output. In order to output location, the scan developer must interrogate GPS+Visual image for each location measurement. This output is then stored into cloud storage 430 measurement. Some outputs require a processed parameter plus multiple raw data inputs. An example of this is soil nutrient and the subset of this being nitrogen. Nitrogen measurement requires a location stamp (data 430) plus spectroscopic information (data 410).

    [0053] After all measured data has been processed and stored it is available for the farm crop yield simulator 450. Data 450 is accessed from the cloud storage 430 plus uses inputs from other sources (climate, farm manager input). Farmer interface 460 directs which software module used by the simulator. Software modules are generated by multiple institutions.

    [0054] FIG. 500 shows the minimum compliment of sensing devices required to scan farmland required by in some embodiments of the system. Two vehicles are introduced: a ground vehicle and an air vehicle. Air vehicles include satellites, drones, helicopters, and fixed wing aircraft. Each equipment set is managed and run using a computer 525/565, which performs many tasks from location determination to scheduling, to uploading data to name a few. The computer all flight information is loaded to the computer prior to flight. Once loaded the computer is capable of executing the measurement plan autonomously. Once the flight is completed, data is delivered to the cloud via the access point 523/563 which can be a wireless technology, a wired technology, or a hand carried storage device.

    [0055] GPS location is determined using GPS sensors 521/561 that connect to the controller 525/565. The compliment set of sensors required to perform farm sensing is also described in FIG. 500 and includes Radar 511/522, and one or multiple spectrometers 515/555 and optical cameras 519/559. Air vehicles may require an additional thermal imager 557. Ground vehicles may require an additional temperature and humidity sensor 517. LiDAR 513/555 may be added if local LiDAR data is not available.

    [0056] The controller computer 525/565 is loaded with a route with locations for measurement. The controller 525/565 measures GPS and uses visual imagery to determine true location. Once the device is at the location required the controller signals measurements from different instruments as a function of their spot size. After each measurement, the controller retrieves the data and time/date/location stamps each measurement and stores it inside local memory. Once information from the access point is ready to upload data, the controller uploads its data and is ready for the next set of measurements.

    [0057] FIG. 600 diagrams the sensors required to generate each parameter required by the simulator. Most sensor systems require multiple inputs from multiple sensors to develop an accurate result. This system is no different and as such we have defined a minimum set of sensors which will achieve our goals of modeling a modern farm. Our current system consists of approximately 6 sensor types which consist of Radar, LiDAR, Spectrometry, Thermal imager, Optical imager, and GPS. Typically, the spectrometer consists of multiple boxes which focus on separate frequencies required to scan the full band. FIG. 600 diagrams which sensors are required for each parameter. It is important to note that ALL sensors minus LiDAR (noting some states have extensive LiDAR maps which invalidate the need to add the sensor) are required to operate the crop yield simulation tool of FIG. 660. Each sensor is discussed below:

    [0058] Location 610—Location requires GPS+Visual imagery as a minimum set of inputs.

    [0059] Soil type 615—Soil Moisture and Soil Type requires radar imagery as a function of time as defined in the reference patents.

    [0060] Microclimates 620—Microclimate Models require input from Location, thermal imagery and External climate models

    [0061] Rainfall model 625—Rainfall Model requires inputs from location, Microclimate model, and NOAA climate forecast

    [0062] Soil nutrient model 630—Soil Nutrient Model f(t)—Requires Location, LiDAR surface contour, Crop Maturity Model, Water transport model

    [0063] Pest infestation 640—Pest infestation scan requires inputs from Location, Visual Imagery, Spectroscopy

    [0064] Chemical application 645—Pest chemical application requires inputs from Location, Microclimates, External climate models.

    [0065] Farm practices optimization 650—Farm Practices Optimization—Requires input from Soil Type and Soil Moisture and LiDAR, and ground surface condition (fanner input).

    [0066] Crop health 655—Crop Health/Stress—Requires input from Location, Spectroscopy, Soil Moisture, Soil Nutrient data base.

    [0067] Crop yield simulator 660—Crop Yield Simulator—requires input from all inputs described in this document.

    [0068] FIG. 700 shows controller alignment of disparate sensors when measuring a volume. FIG. 700 shows the vehicle direction and location 750. On board calculation of location is performed by the sensor computer FIG. 525/565. The computer calculates when to enable each of the sensors that it controls. The computer first takes into account the area of interest 760 then, knowing each sensor focal length and spot size the computer plans a scan for each of the multiple on board sensors and as the vehicle comes into position executes the plan. An example of this is to look at three spot sizes 740 shown in FIG. 700. The largest spot size are the visual and thermal sensor spot sizes 310. This requires only one picture in the area of interest 760. The synthetic aperture radar spot size 730 at this altitude requires 9 measurements as shown in FIG. 700 and therefore the controller schedules nine per area of interest 760. The last sensor is the LiDAR 740, which is a scan even though; the LiDAR requires start stop timing.

    [0069] Focal point calculations are required for the radar as a function of aircraft height and therefore the controller has to adjust number of spots proportionally to distance from target. For instance, if the distance is 500 feet in altitude the radar measures 49 spots, if the altitude is 1000 feet the radar must be set up to measure 9 spots.

    [0070] FIG. 800 illustrates an optical instrument stabilization and housing design. Optical instruments used in conjunction with other measurements must be stabilized from vibration to minimize/eliminate blur caused by vibration of vehicle in motion. FIG. 800 shows a simplified version of the design however this design extends to mounting to all devices individually or separate onto a stabilization table 820 which uses multiple passive or active stabilization feet 230.

    [0071] Imaging equipment is typically also sensitive to dust so the enclosure introduced in FIG. 800 includes the use of an optically transparent cover which is attached to the enclosure FIG. 250 such that no air or water or dust might leak in.

    [0072] Custom lenses 840 are specially adapted to the imager components 210. These lenses allow for adjustment of focal points such that each piece of sensing equipment might focus at the proper range and exhibit a designed for spot size.

    [0073] FIG. 900 shows the position vector error correction methodology utilizing two sensors referenced herein. This methodology requires a reference location image 910 stored prior to scanning and a photograph of the current aircraft location to be taken when the vehicle is located within range of the GPS target location 920. The system then correlates one spot or one surface on the reference map to one spot or one surface on the current measured picture and generates two vectors Vr and Vp in three-dimensional space. The error vector Verr is the difference between the two.

    [0074] Once the difference vector 940 is found, all location measurements are adjusted by subtracting the error vector 940 from the GPS location vector.

    [0075] While particular preferred and alternative embodiments of the present intention have been disclosed, it will be appreciated that many various modifications and extensions of the above described technology may be implemented using the teaching of this invention. All such modifications and extensions are intended to be included within the true spirit and scope of the appended claims.