Patent classifications
G06T7/285
IMAGE PROCESSING-BASED WEIGHT ESTIMATION FOR AQUACULTURE
Methods, systems, and apparatus, including computer programs encoded on computer-storage media, for fish weight estimation based on fish tracks identified in images. In some implementations, a method includes obtaining images of fish enclosed in a fish enclosure, identifying fish tracks shown in the images of the fish, determining a quality score for each of the fish tracks, selecting a subset of the fish tracks based on the quality scores, determining a representative weight of the fish in the fish enclosure based on weights of the fish shown in the subset of the fish tracks, and outputting the representative weight for display or storage at a device connected to the one or more processors.
Systems and methods for navigating a vehicle among encroaching vehicles
Systems and methods use cameras to provide autonomous navigation features. In one implementation, a method for navigating a user vehicle may include acquiring, using at least one image capture device, a plurality of images of an area in a vicinity of the user vehicle; determining from the plurality of images a first lane constraint on a first side of the user vehicle and a second lane constraint on a second side of the user vehicle opposite to the first side of the user vehicle; enabling the user vehicle to pass a target vehicle if the target vehicle is determined to be in a lane different from the lane in which the user vehicle is traveling; and causing the user vehicle to abort the pass before completion of the pass, if the target vehicle is determined to be entering the lane in which the user vehicle is traveling.
Systems and methods for navigating a vehicle among encroaching vehicles
Systems and methods use cameras to provide autonomous navigation features. In one implementation, a method for navigating a user vehicle may include acquiring, using at least one image capture device, a plurality of images of an area in a vicinity of the user vehicle; determining from the plurality of images a first lane constraint on a first side of the user vehicle and a second lane constraint on a second side of the user vehicle opposite to the first side of the user vehicle; enabling the user vehicle to pass a target vehicle if the target vehicle is determined to be in a lane different from the lane in which the user vehicle is traveling; and causing the user vehicle to abort the pass before completion of the pass, if the target vehicle is determined to be entering the lane in which the user vehicle is traveling.
STEREO CAMERA APPARATUS AND CONTROL DEVICE
A stereo camera apparatus includes a stereo camera, a speed sensor, and a control device. The control device includes one or more processors and one or more storage media. The one or more processors are configured to: detect corresponding points from a first image pair and a second image pair to be captured at different times by the stereo camera; divide each of images of the first image pair and the second image pair into regions; calculate, for each of the regions, a movement speed based on an external parameter by using one or more of the corresponding points included in the each of the regions; and calculate, for the each of the regions, a parallax correction value to cause a difference between the movement speed based on the external parameter and a movement speed detectable by the speed sensor to fall below a predetermined threshold.
STEREO CAMERA APPARATUS AND CONTROL DEVICE
A stereo camera apparatus includes a stereo camera, a speed sensor, and a control device. The control device includes one or more processors and one or more storage media. The one or more processors are configured to: detect corresponding points from a first image pair and a second image pair to be captured at different times by the stereo camera; divide each of images of the first image pair and the second image pair into regions; calculate, for each of the regions, a movement speed based on an external parameter by using one or more of the corresponding points included in the each of the regions; and calculate, for the each of the regions, a parallax correction value to cause a difference between the movement speed based on the external parameter and a movement speed detectable by the speed sensor to fall below a predetermined threshold.
SYSTEM AND METHOD FOR GENERATING COMBINED EMBEDDED MULTI-VIEW INTERACTIVE DIGITAL MEDIA REPRESENTATIONS
Various embodiments describe systems and processes for capturing and generating multi-view interactive digital media representations (MIDMRs). In one aspect, a method for automatically generating a MIDMR comprises obtaining a first MIDMR and a second MIDMR. The first MIDMR includes a convex or concave motion capture using a recording device and is a general object MIDMR. The second MIDMR is a specific feature MIDMR. The first and second MIDMRs may be obtained using different capture motions. A third MIDMR is generated from the first and second MIDMRs, and is a combined embedded MIDMR. The combined embedded MIDMR may comprise the second MIDMR being embedded in the first MIDMR, forming an embedded second MIDMR. The third MIDMR may include a general view in which the first MIDMR is displayed for interactive viewing by a user on a user device. The embedded second MIDMR may not be viewable in the general view.
SYSTEM AND METHOD FOR GENERATING COMBINED EMBEDDED MULTI-VIEW INTERACTIVE DIGITAL MEDIA REPRESENTATIONS
Various embodiments describe systems and processes for capturing and generating multi-view interactive digital media representations (MIDMRs). In one aspect, a method for automatically generating a MIDMR comprises obtaining a first MIDMR and a second MIDMR. The first MIDMR includes a convex or concave motion capture using a recording device and is a general object MIDMR. The second MIDMR is a specific feature MIDMR. The first and second MIDMRs may be obtained using different capture motions. A third MIDMR is generated from the first and second MIDMRs, and is a combined embedded MIDMR. The combined embedded MIDMR may comprise the second MIDMR being embedded in the first MIDMR, forming an embedded second MIDMR. The third MIDMR may include a general view in which the first MIDMR is displayed for interactive viewing by a user on a user device. The embedded second MIDMR may not be viewable in the general view.
Binocular-vision-based method for tracking fruit space attitude and fruit space motion
A binocular-vision-based method for tracking fruit space attitude and fruit space motion, the method comprising: establishing a connected base coordinate system by taking a junction of a fruit and a fruit stem as an origin; statically photographing a feature point on the surface of the fruit and a point of the connected base coordinate system established at the junction of the fruit and the fruit stem; storing a photographed image; acquiring an inherent relationship between the feature point and the connected base coordinate system; photographing dynamic motion of the fruit; acquiring absolute coordinates of the feature point on the surface of the fruit; calculating, according to the inherent relationship between the feature point and the connected base coordinate system, absolute coordinates of a point of the connected base coordinate system at each moment corresponding to each frame of image; and respectively calculating the displacement, instantaneous speed and instantaneous acceleration of the fruit, calculating swing angular displacement and swing angular acceleration of the fruit, and calculating a fruit torsion angular speed and a fruit torsion angular acceleration at the moment t. The study of a fruit motion state in the field of forest fruit harvest through vibration is performed, so that the motion of fruits can be better tracked.
Binocular-vision-based method for tracking fruit space attitude and fruit space motion
A binocular-vision-based method for tracking fruit space attitude and fruit space motion, the method comprising: establishing a connected base coordinate system by taking a junction of a fruit and a fruit stem as an origin; statically photographing a feature point on the surface of the fruit and a point of the connected base coordinate system established at the junction of the fruit and the fruit stem; storing a photographed image; acquiring an inherent relationship between the feature point and the connected base coordinate system; photographing dynamic motion of the fruit; acquiring absolute coordinates of the feature point on the surface of the fruit; calculating, according to the inherent relationship between the feature point and the connected base coordinate system, absolute coordinates of a point of the connected base coordinate system at each moment corresponding to each frame of image; and respectively calculating the displacement, instantaneous speed and instantaneous acceleration of the fruit, calculating swing angular displacement and swing angular acceleration of the fruit, and calculating a fruit torsion angular speed and a fruit torsion angular acceleration at the moment t. The study of a fruit motion state in the field of forest fruit harvest through vibration is performed, so that the motion of fruits can be better tracked.
Cascaded architecture for disparity and motion prediction with block matching and convolutional neural network (CNN)
A CNN operates on the disparity or motion outputs of a block matching hardware module, such as a DMPAC module, to produce refined disparity or motion streams which improve operations in images having ambiguous regions. As the block matching hardware module provides most of the processing, the CNN can be small and thus able to operate in real time, in contrast to CNNs which are performing all of the processing. In one example, the CNN operation is performed only if the block hardware module output confidence level is below a predetermined amount. The CNN can have a number of different configurations and still be sufficiently small to operate in real time on conventional platforms.