METHOD AND DEVICE FOR CROP CANOPY CHLOROPHYLL FLUORESCENCE THREE-DIMENSIONAL DISTRIBUTION INFORMATION ACQUISITION
20220146428 · 2022-05-12
Assignee
Inventors
- Jizhang Wang (Zhenjiang, CN)
- Rongrong Gu (Zhenjiang, CN)
- Yun Zhang (Zhenjiang, CN)
- Junjie Yuan (Zhenjiang, CN)
- Pingping Li (Zhenjiang, CN)
Cpc classification
G01N2021/8466
PHYSICS
G01N21/6486
PHYSICS
International classification
G01B11/00
PHYSICS
Abstract
A method and a device for crop canopy chlorophyll fluorescence three-dimensional distribution information acquisition are provided. The device includes a Cropobserver canopy chlorophyll fluorescence detection device, a 3D camera and a computer system. The 3D camera is connected to the computer system, Visual studio 2017 and MATLAB 2018 are run in the computer system, and the Visual studio 2017 calls a point cloud library and a computer vision library to realize three-dimensional visualization of chlorophyll fluorescence information of crops to be tested. By means of the new method and the new device, the problem of incompleteness of the two-dimensional chlorophyll fluorescence information distribution acquired is solved, overall 3D visual distribution of crop canopy chlorophyll fluorescence distribution is realized, and important technical support is provided for acquisition and research of three-dimensional visual distribution information of chlorophyll fluorescence of the whole crop canopy.
Claims
1. A method for crop canopy chlorophyll fluorescence three-dimensional (3D) distribution information acquisition, comprising: respectively obtaining depth images and mapped color images of laser dots emitted by a fluorescence-induced laser emitter on a background plate before and after raising by using a 3D camera, and calibrating the depth images and the mapped color images to obtain camera intrinsic matrices; obtaining spatial coordinates of the laser dots based on pixel coordinates of edge points in the color images and depth values of the edge points in the depth images in combination with the camera intrinsic matrices; obtaining a spatial linear equation according to the spatial coordinates of the laser dots, and solving spatial coordinate O.sub.2(a, b, c) of an aperture center of the fluorescence-induced laser emitter relative to the 3D camera; acquiring chlorophyll fluorescence information of a crop canopy to be tested, mapping dot sequence number coordinates (g′, h′) of effective chlorophyll fluorescence signals to pixel coordinates in the color images and depth information (u′″, v′″, d′″) of the depth images; correspondingly characterizing (u′″, v′″, d′″) to spatial coordinates (x′, y′, z′) using an aperture center of a depth sensor in the 3D camera as a spatial coordinate origin, and correspondingly characterizing a chlorophyll fluorescence information signal sequence of the crop canopy to be tested to spatial coordinates (x′+a, y′+b, z′+c) using the aperture center of the fluorescence-induced laser emitter as a spatial coordinate origin; performing three-dimensional visualization of the chlorophyll fluorescence information of the crop canopy to be tested; respectively generating, based on data in Text4-Text6 by using a pointcloud function for point cloud generation, point clouds pointcloud-Yield-Kinect, pointcloud-PAR-Kinect, and pointcloud-rETR-Kinect that comprise spatial coordinates and chlorophyll fluorescence information and use the aperture center of the depth sensor as an origin; respectively generating, based on data in Text7-Text9 by using the pointcloud function for point cloud generation, point clouds pointcloud-Yield-CropObserver, pointcloud-PAR-CropObserver, and pointcloud-rETR-CropObserver that comprise spatial coordinates and chlorophyll fluorescence information and use the aperture center of the fluorescence-induced laser emitter as an origin; wherein the Text4 comprises data
2. The method for crop canopy chlorophyll fluorescence three-dimensional distribution information acquisition according to claim 1, wherein the spatial coordinates of the laser dots are expressed as M(x, y, z), and
3. The method for crop canopy chlorophyll fluorescence three-dimensional distribution information acquisition according to claim 1, wherein the
4. The method for crop canopy chlorophyll fluorescence three-dimensional distribution information acquisition according to claim 1, wherein dot sequence numbers (g, h) corresponding to pixel coordinates (u.sub.A1, v.sub.A1), (u.sub.B1, v.sub.B1), (u.sub.C1, v.sub.C1), and (u.sub.D1, v.sub.D1) of edge points of the mapped color images are respectively (1, 1), (e, 1), (e, f), and (1, f), and pixel coordinates, which are corresponding to the dot sequence numbers (g, h), in the depth images captured by the 3D camera are recorded as points (u″, v″), wherein u″=(g−1)Δ.sub.x+u.sub.D1, and v″=(h−1)Δ.sub.y+v.sub.D1, wherein e is a number of dots generated by a canopy chlorophyll fluorescence detection device in a row direction, and f is a number of dots generated by the canopy chlorophyll fluorescence detection device in a column direction.
5. The method for crop canopy chlorophyll fluorescence three-dimensional distribution information acquisition according to claim 4, wherein
u′″=(g′−1)Δ.sub.x+u.sub.D1,v′″=(h′−1)Δ.sub.y+v.sub.D1, wherein Δ.sub.x is a pixel distance between neighboring dots generated by the fluorescence-induced laser emitter in the row direction, and Δ.sub.y is a pixel distance between neighboring dots generated by the fluorescence-induced laser emitter in the column direction.
6. The method for crop canopy chlorophyll fluorescence three-dimensional distribution information acquisition according to claim 5, wherein
7. The method for crop canopy chlorophyll fluorescence three-dimensional distribution information acquisition according to claim 6, wherein
8. The method for crop canopy chlorophyll fluorescence three-dimensional distribution information acquisition according to claim 6, wherein
9. The method for crop canopy chlorophyll fluorescence three-dimensional distribution information acquisition according to claim 1, further comprising: acquiring canopy chlorophyll fluorescence three-dimensional distribution information of different growth sequences of crops to be tested.
10. A device for crop canopy chlorophyll fluorescence three-dimensional distribution information acquisition for the method according to claim 1, comprising: canopy chlorophyll fluorescence detection device, a 3D camera and a computer system, wherein the 3D camera is connected to the computer system, Visual studio 2017 and MATLAB 2018 are run in the computer system, and the Visual studio 2017 calls a point cloud library and a computer vision library to realize three-dimensional visualization of chlorophyll fluorescence information of crops to be tested.
11. The device according to claim 10, wherein the spatial coordinates of the laser dots are expressed as M(x, y, z), and
12. The device according to claim 10, wherein the
13. The device according to claim 10, wherein dot sequence numbers (g, h) corresponding to pixel coordinates (u.sub.A1, v.sub.A1), (u.sub.B1, v.sub.B1), (u.sub.C1, v.sub.C1), and (u.sub.D1, v.sub.D1) of edge points of the mapped color images are respectively (1, 1), (e, 1), (e, f), and (1, f), and pixel coordinates, which are corresponding to the dot sequence numbers (g, h), in the depth images captured by the 3D camera are recorded as points (u″, v″), wherein u″=(g−1)Δ.sub.x+u.sub.D1, and v″=(h−1)Δ.sub.y+v.sub.D1, wherein e is a number of dots generated by a canopy chlorophyll fluorescence detection device in a row direction, and f is a number of dots generated by the canopy chlorophyll fluorescence detection device in a column direction.
14. The device according to claim 13, wherein
u′″=(g′−1)Δ.sub.x+u.sub.D1,v′″=(h′−1)Δ.sub.y+v.sub.D1, wherein Δ.sub.x is a pixel distance between neighboring dots generated by the fluorescence-induced laser emitter in the row direction, and Δ.sub.y is a pixel distance between neighboring dots generated by the fluorescence-induced laser emitter in the column direction.
15. The device according to claim 14, wherein
16. The device according to claim 15, wherein
17. The device according to claim 15, wherein
18. The device according to claim 10, further comprising: acquiring canopy chlorophyll fluorescence three-dimensional distribution information of different growth sequences of crops to be tested.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0028]
[0029]
[0030]
[0031]
[0032] In the drawings: 1—canopy chlorophyll fluorescence detection device, 1-1—fluorescence-induced laser emitter, 1-2—LI-COR optical quantum sensor, 1-3—chlorophyll fluorescence sensor, 1-4—HDMI port, 1-5—24V power input port, 1-6—USB3.0 port, 1-7—voltage converter, 1-8—iron chain, 2—3D camera, 3—triangular support, 4—computer system, 5—display, 6—mobile storage device, 7—crop to be tested, 8—black background plate, 9—movable rack, 10—universal wheel.
DETAILED DESCRIPTION OF THE EMBODIMENTS
[0033] The present invention will be further described in detail below with reference to drawings and embodiments, but the protection scope of the present invention is not limited thereto.
[0034] As shown in
[0035] In this example, the crop to be tested 7 is cucumber, the canopy chlorophyll fluorescence detection device 1 is the Cropobserver canopy chlorophyll fluorescence detection device manufactured by Phenotrait, the Netherlands, the 3D camera 2 is Microsoft's Kinect V2 depth camera, and the computer system 4 is Windows 10 system. The information acquisition control function of the 3D camera 2 implements the acquisition of color images and depth images in Visual studio 2017, the calibration function of the 3D camera 2 is implemented in MATLAB 2018 by a checkerboard calibration kit based on the principle of Zhengyou Zhang's calibration, and the point cloud acquisition function and the chlorophyll fluorescence information characterization visualization function of the 3D camera 2 are implemented by calling a point cloud library (PCL) and a computer vision library (Open Source Computer Vision Library, OpenCV) in Visual studio 2017. Visual studio 2017 and MATLAB 2018 are software running in the computer system 4.
[0036] As shown in
[0037] Step 1. The canopy chlorophyll fluorescence detection device 1 is disposed.
[0038] The movable rack 9 carries the canopy chlorophyll fluorescence detection device 1 to move to the top of the crop canopy to be tested 7. After initialization of the measurement device, “Centre” and “Test Meas” buttons in the setting interface are successively pressed, to cause laser dots emitted from the fluorescence-induced laser emitter 1-1 to point to the center of the crop to be tested 7. A measurement range is set, so that a range of dots generated by the fluorescence-induced laser emitter 1-1 surrounds the crop to be tested 7. As shown in
[0039] Step 2. Image acquisition and calibration of the 3D camera 2.
[0040] In this example, the 3D camera 2 has a color sensor resolution of 1920×1080, and a depth sensor resolution of 512×424. Visual studio 2017 is run in the computer system 4, and a computer vision library (Open Source Computer Vision Library, OpenCV) and a camera SDK are called, to respectively acquire depth frames data to arrays (DepthFrameDate) and color frames data to arrays (ColorSpacePoint), and respectively saves them as a depth image and a color image. A MapDepthFrameToColorSpace( ) function based on the principle of bilinear interpolation is used to calculate a mapping relationship between the depth image and the color image according to depth frame information, and pixel coordinates in the depth image are mapped to the color image, so that coordinates of pixels in the depth image are mapped to coordinates in the color image, to obtain a 512*424 array. Elements of the array are coordinates in the color image that correspond to the depth image, and contain color information. The array is saved as a mapped color image.
[0041] Mapped color images corresponding to different positions of a checkerboard calibration plate are acquired. The mapped color images are input to Zhengyou Zhang's calibration toolkit in MATLAB 2018. A corner distance of the checkerboard is input. Valid calibration pictures are screened to obtain n checkerboard images with a calibration plot error of less than 0.2 pixels for calibration, where n>20. Then a camera intrinsic matrix is exported:
where f.sub.x=f*sx,f.sub.y=f*sy,f is the focal length of the camera (measured in mm), [sx, sy] represents the number of pixels per millimeter in the (x, y) direction, f.sub.x and f.sub.y respectively represent focal lengths of the camera on the x axis and the y axis (measured in pixels), and [c.sub.x, c.sub.y] is the aperture center of the camera.
[0042] Step 3. Calibration point information capture between the canopy chlorophyll fluorescence detection device 1 and the 3D camera.
[0043] As shown in
[0044] Step 4. World coordinates of calibration edge points are extracted.
[0045] A model structure of the 3D camera 2 is as shown in
[0046] where (u.sub.0, v.sub.0) is the pixel coordinates of the aperture center of the camera; and s is a scaling factor, i.e., a ratio of the depth value to an actual application, and s is generally set to 1000.
[0047] A back calculation formula (2) may be written as follows: when a point m(u, v, d) is known, a corresponding spatial coordinate point M(x, y, z) is derived:
[0048] The depth images and the mapped color images of the four red laser dots A.sub.1, B.sub.1, C.sub.1, and D.sub.1 at the edge and the four red laser dots A.sub.2, B.sub.2, C.sub.2, and D.sub.2 at the edge that are acquired in step 3 are imported into Matlab 2018. For the color image, the color image is grayed using a super red grayscale factor 2R-G-B (where R, G, and B are three color channel components: red, green, and blue), to acquire red characteristics of the laser dots at the edge, and obtain clear edge points. The pixel coordinates of the edge points are extracted: (u.sub.A1, v.sub.A1), (u.sub.B1, v.sub.B1), . . . , and (u.sub.D2, v.sub.D2). (u.sub.A1, v.sub.A1), (u.sub.B1, v.sub.B1), . . . , and (u.sub.D2, v.sub.D2) are mapped to the depth image, to obtain depths d.sub.A1, d.sub.B1, . . . , and d.sub.D2. With reference to the intrinsic matrices Intrinsic Matrix-a and Intrinsic Matrix-b in step 3, spatial coordinate points of A.sub.1, B.sub.1, C.sub.1, D.sub.1, A.sub.2, B.sub.2, C.sub.2, and D.sub.2 are acquired: (x.sub.A1, y.sub.A1, z.sub.A1), (x.sub.B1, y.sub.B1, z.sub.B1), . . . , and (x.sub.D2, y.sub.D2, z.sub.D2).
[0049] Step 5. A spatial position of an aperture center O.sub.2 of the fluorescence-induced laser emitter relative to the 3D camera 2 is calibrated.
[0050] According to the coordinates of the spatial coordinate points A.sub.1, B.sub.1, C.sub.1, D.sub.1, A.sub.2, B.sub.2, C.sub.2, and D.sub.2 acquired in step 4, spatial linear equations passing through A.sub.1 A.sub.2, B.sub.1 B.sub.2, C.sub.1 C.sub.2, and D.sub.1 D.sub.2 are set up, which are respectively recorded as straight lines l.sub.1, l.sub.2, l.sub.3, and l.sub.4. Assume that the linear equations areas follows:
[0051] In the above linear equations, N.sub.1, N.sub.2, N.sub.3, and N.sub.4 are constants. By substituting points A.sub.1, B.sub.1, C.sub.1, and D.sub.1 into the straight lines l.sub.1, l.sub.2, l.sub.3, and l.sub.4 respectively, the constants N.sub.1, N.sub.2, N.sub.3, and N.sub.4 can be calculated, and then the four linear equations l.sub.1, l.sub.2, l.sub.3, and l.sub.4 can be solved.
[0052] Because the points A.sub.1, B.sub.1, C.sub.1, D.sub.1, A.sub.2, B.sub.2, C.sub.2, and D.sub.2 are emitted from the aperture center O.sub.2 of the fluorescence-induced laser emitter, all the straight lines l.sub.1, l.sub.2, l.sub.3, and l.sub.4 pass through the point O.sub.2. It is assumed that the spatial coordinates of the aperture center of the fluorescence-induced laser emitter relative to the camera are O.sub.2(a, b, c), including three unknowns a, b, and c. By substitution into any three of the above linear equations, the spatial coordinates O.sub.2(a, b, c) of the aperture center of the fluorescence-induced laser emitter relative to the camera can be calculated.
[0053] Step 6. The canopy chlorophyll fluorescence detection device 1 acquires chlorophyll fluorescence information of the crop to be tested.
[0054] The number of dots generated by the Cropobserver in the x-axis direction is set to e=50, and that in the y-axis direction is set to f=50, a boundary for dots generated by the fluorescence-induced laser emitter 1-1 is A.sub.1, B.sub.1, C.sub.1, and D.sub.1, and a 50×50 dot array is formed, where sequence numbers in the array are recorded as (g, h) (1≤g≤50, 1≤h≤50). Neighboring dots are spaced from each other in the row direction by the same distance, which is recorded as a constant n.sub.1. Neighboring dots are spaced from each other in the column direction by the same distance, which is recorded as a constant n.sub.2. The value of
equals to a ratio
between scanning angle ratios in the row direction and the column direction. “Start scan” is clicked to start measurement. The fluorescence-induced laser emitter 1-1 generates dots in the following order: first generating a dot at point A.sub.1, the sequence number of the dot being recorded as (1, 1); generating 49 dots in a direction toward point D.sub.1 at equal intervals of n.sub.1, the sequence number of the dot at point D.sub.1 being recorded as (1, 50); then generating a dot at a position that is distant from point A.sub.1 downward by a distance n.sub.2, the sequence number of the dot being recorded as (2, 1); then generating 49 dots in a direction toward point D.sub.1 downward by a distance n.sub.2, the sequence number of the dot that is distant from D.sub.1 downward by n.sub.2 being recorded as (2, 50); then generating a dot at a position that is distant from point A.sub.1 downward by a distance 2n.sub.2, the sequence number of the dot being recorded as (3, 1), and then generating 49 dots in a direction toward point D.sub.1 downward by a distance 2n.sub.2 at intervals of n.sub.1, the sequence number of the dot that is distant from D.sub.1 downward by 2n.sub.2 being recorded as (3, 50); and so on. Dots are generated in sequence based on the above rule, the sequence number of the dot at point B.sub.1 being recorded as (50, 1). Finally, a dot is generated at point C.sub.1, the sequence number of the dot at point C, being recorded as (50, 50). The position of the measurement point is changed every 5 seconds, and the chlorophyll fluorescence sensor 1-3 acquires and stores a position at which the crop to be tested 7 reflects chlorophyll fluorescence and fluorescence data of this position.
[0055] The canopy chlorophyll fluorescence detection device 1 mainly measures the following parameters: (1) photochemical efficiency: maximum photochemical efficiency
of leaves under dark adaptation, and actual photochemical efficiency
of leaves under light adaptation; (2) PAR: relative light intensity on the leaf surface; (3) rETR: relative electron transfer rate in leaves. Fv=Fm−F.sub.0, where Fm is maximum chlorophyll fluorescence measured under dark adaptation conditions, and F.sub.0 is an initial value of the chlorophyll fluorescence parameter measured under dark adaptation conditions; F.sub.q′=F.sub.m′−F.sub.t, where F.sub.m′ is maximum fluorescence under light adaptation, i.e., a fluorescence intensity when all PSII reaction centers are closed under light adaptation, and F.sub.t is real-time fluorescence of the crop after receiving light for a period of time t; the relative electron transfer rate rETR=0.425×(F.sub.q′/F.sub.m′)×PAR. When the canopy chlorophyll fluorescence detection device 1 operates, the computer system 4 captures depth images and mapped color images of the crop to be tested 7 using the 3D camera 2, where the depth images including pixel and depth information are expressed as (u′,v′,d′), and the color images including three color channels, red r′, green g′, and blue b′, are expressed as (u′,v′,r′,g′,b′).
[0056] Step 7. The depth images and the mapped color images of the crop to be tested are converted to point clouds for displaying.
[0057] A point cloud library (PCL) and a computer vision library (Open Source Computer Vision Library, OpenCV) are called in Visual studio 2017, and by traversing (u′,v′,d′) acquired in step 6, the crop depth images are converted based on formula (3) into spatial coordinate points (X, Y, Z), which are saved in a matrix XYZ of three columns, respectively named X, Y, and Z. The three color channel components red, green, and blue of (r′, g′, b′) acquired in step 6 are separated to form three channel components r, g, and b, which are respectively saved in a matrix RGB of three columns, respectively named R, G, and B. Point cloud plots are generated from the matrix components X, Y, Z, R, G, and B by using a pointcloud function for point cloud generation.
[0058] Step 8. Segmentation is performed for the crop canopy to be tested.
[0059] The point cloud plots in step 7 also contain background point cloud information in addition to the crop to be tested. The point cloud plots in step 6 are processed using a super green grayscale operation (2R-G-B), to highlight the green crop point cloud part. A binarization thresholding operator THRESH_OTSU in OpenCV is used for thresholding, to separate the green crop point cloud.
[0060] Step 9. Dot sequence number coordinates of effective chlorophyll fluorescence signals are mapped to pixel coordinates in the depth images and the mapped color images.
[0061] The sequence numbers (g, h) of the dots generated by the canopy chlorophyll fluorescence detection device 1 in step 6 corresponding to the pixel coordinates (u.sub.A1, v.sub.A1), (u.sub.B1, v.sub.B1), (u.sub.C1, v.sub.C1), and (u.sub.D1, v.sub.D1) of the edge points in step 4 are respectively (1, 1), (e, 1), (e, f), and (1, f), and pixel pitches p.sub.x and p.sub.y corresponding to head-to-tail distances between the dots generated by the canopy chlorophyll fluorescence detection device 1 in the row direction and the column direction are respectively:
where the dot sequence numbers (g, h) are evenly distributed in the pixel coordinate plane. The pixel pitch between neighboring dots generated by the fluorescence-induced laser emitter in the row direction is recorded as Δ.sub.x, and the pixel pitch between neighboring dots generated by the fluorescence-induced laser emitter in the column direction is recorded as Δ.sub.y. Assuming that a dot array generated by the fluorescence-induced laser emitter is
The pixel coordinates in the depth images captured by the camera corresponding to the dot sequence numbers (g, h) are recorded as points (u″, v″), where u″=(g−1)Δ.sub.x+u.sub.D1, and v″=(h−1)Δ.sub.y+V.sub.D1. A sequence number of a dot with a chlorophyll fluorescence signal recorded in a cycle of the canopy chlorophyll fluorescence detection device 1 is found. Coordinate information and chlorophyll fluorescence information of this sequence number are sequentially saved in a row in Text1 in the following order:
The depth images corresponding to (g′, h′) are (u′″,v′″,d′″), u′″=(g′−1)Δ.sub.x+u.sub.D1, and v′″=(h′−1)Δ.sub.y+v.sub.D1. The pixel coordinates, depth information and chlorophyll fluorescence information of the depth image are sequentially saved in a row in Text2 in the following order:
[0062] Step 10. A chlorophyll fluorescence information signal sequence of the crop canopy to be tested are correspondingly characterized to spatial coordinates using the aperture center of the depth sensor as a spatial coordinate origin.
[0063] Based on the coordinate conversion method in formula (3), the first three columns of pixel and depth coordinates (u′″, v′″, d′″) in Text2 are converted into spatial coordinates (x′, y′, z′) using the aperture center of the depth sensor as a spatial coordinate origin, which are sequentially saved, together with the last three columns in Text2, in a row in Text3 in the following order:
[0064] Step 11. A chlorophyll fluorescence information signal sequence of the crop canopy to be tested are correspondingly characterized to spatial coordinates using the aperture center of the fluorescence-induced laser emitter as a spatial coordinate origin.
[0065] According to the spatial coordinates (x′, y′, z′) using the aperture center of the depth sensor as the spatial coordinate origin in step 10, the spatial coordinates O.sub.2(a, b, c) of the aperture center of the fluorescence-induced laser emitter relative to the camera have been obtained in step 5, and thus spatial coordinates of effective chlorophyll fluorescence signals using the aperture center of the fluorescence-induced laser emitter as the origin of space are (x′+a, y′+b, z′+c).
[0066] Step 12. Three-dimensional visualization of the canopy chlorophyll fluorescence information of the crop to be tested is performed.
[0067] The last three columns of data in Text3 are converted into
and rETR, so that their value ranges are 0 to 255, i.e., the three columns of chlorophyll fluorescence information data fall within value ranges of the red, green, and blue color channels. Data is sequentially saved in a row in Text4 in the following order: x′, y′, z′,
Data is sequentially saved in a row in Text5 in the following order: x′, y′, z′, 0, PAR/10, 0. Data is sequentially saved in a row in Text6 in the following order: x′, y′, z′, 0, 0, rETR. Data is sequentially saved in a row in Text7 in the following order:
Data is sequentially saved in a row in Text8 in the following order: x′+a, y′+b, z′+c, 0, PAR/10, 0. Data is sequentially saved in a row in Text9 in the following order: x′+a, y′+b, z′+c, 0, 0, rETR. The PCL and the OpenCV are called in Visual studio 2017. Point clouds pointcloud-Yield-Kinect, pointcloud-PAR-Kinect, and pointcloud-rETR-Kinect that include spatial coordinates and chlorophyll fluorescence information and use the aperture center of the depth sensor as an origin are respectively generated based on data in Text4-Text6 by using a pointcloud function for point cloud generation. Point clouds pointcloud-Yield-CropObserver, pointcloud-PAR-CropObserver, and pointcloud-rETR-CropObserver that include spatial coordinates and chlorophyll fluorescence information and use the aperture center of the fluorescence-induced laser emitter as an origin are respectively generated based on data in Text7-Text9 by using a pointcloud function for point cloud generation.
[0068] The point clouds including the spatial coordinates and the chlorophyll fluorescence information are characterized to the green crop point cloud separated in step 8 by using a pcshowpair( ) function, to form a three-dimensional visual distribution of the chlorophyll fluorescence information on the plant.
[0069] Step 13. Crop canopy chlorophyll fluorescence three-dimensional point cloud distribution information is acquired for different growth sequences of the crop to be tested. In this example, the growth of the cucumber crop is divided into a germination period, a seedling period, a flowering period and a fruiting period, and three-dimensional point cloud distribution information of the chlorophyll fluorescence of the crop canopy to be tested is acquired by performing steps 1 to 12.
[0070] The embodiments are exemplary embodiments of the present invention, but the present invention is not limited to the above-mentioned embodiments. Any obvious improvement, replacement or variation that can be made by one skilled in the art without departing from the spirit of the present invention belongs to the protection scope of the present invention.