Patent classifications
G06T7/11
Global and local binary pattern image crack segmentation method based on robot vision
A global and local binary pattern image crack segmentation method based on robot vision comprises the following steps: enhancing a contrast of an acquired original image to obtain an enhanced map; using an improved local binary pattern detection algorithm to process the enhanced map and construct a saliency map; using the enhanced map and the saliency map to segment cracks and obtaining a global and local binary pattern automatic crack segmentation method; and evaluating performance of the obtained global and local binary pattern automatic crack segmentation method. The present application uses logarithmic transformation to enhance the contrast of a crack image, so that information of dark parts of the cracks is richer. Texture features of a rotation invariant local binary pattern are improved. Global information of four directions is integrated, and the law of universal gravitation and gray and roundness features are introduced to correct crack segmentation results, thereby improving segmentation accuracy. Crack regions can be segmented in the background of uneven illumination and complex textures. The method has good robustness and meets requirements of online detection.
High efficiency dynamic contrast processing
A high efficiency method of processing images to provide perceptual high-contrast output. Pixel intensities are calculated by a weighted combination of a fixed number of static bounding rectangle sizes. This is more performant than incrementally growing the bounding rectangle size and performing expensive analysis on resultant histograms. To mitigate image artifacts and noise, blurring and down-sampling are applied to the image prior to processing.
High efficiency dynamic contrast processing
A high efficiency method of processing images to provide perceptual high-contrast output. Pixel intensities are calculated by a weighted combination of a fixed number of static bounding rectangle sizes. This is more performant than incrementally growing the bounding rectangle size and performing expensive analysis on resultant histograms. To mitigate image artifacts and noise, blurring and down-sampling are applied to the image prior to processing.
Scene-aware video encoder system and method
Embodiments of the present disclosure discloses a scene-aware video encoder system. The scene-aware encoder system transforms a sequence of video frames of a video of a scene into a spatio-temporal scene graph. The spatio-temporal scene graph includes nodes representing one or multiple static and dynamic objects in the scene. Each node of the spatio-temporal scene graph describes an appearance, a location, and/or a motion of each of the objects (static and dynamic objects) at different time instances. The nodes of the spatio-temporal scene graph are embedded into a latent space using a spatio-temporal transformer encoding different combinations of different nodes of the spatio-temporal scene graph corresponding to different spatio-temporal volumes of the scene. Each node of the different nodes encoded in each of the combinations is weighted with an attention score determined as a function of similarities of spatio-temporal locations of the different nodes in the combination.
Scene-aware video encoder system and method
Embodiments of the present disclosure discloses a scene-aware video encoder system. The scene-aware encoder system transforms a sequence of video frames of a video of a scene into a spatio-temporal scene graph. The spatio-temporal scene graph includes nodes representing one or multiple static and dynamic objects in the scene. Each node of the spatio-temporal scene graph describes an appearance, a location, and/or a motion of each of the objects (static and dynamic objects) at different time instances. The nodes of the spatio-temporal scene graph are embedded into a latent space using a spatio-temporal transformer encoding different combinations of different nodes of the spatio-temporal scene graph corresponding to different spatio-temporal volumes of the scene. Each node of the different nodes encoded in each of the combinations is weighted with an attention score determined as a function of similarities of spatio-temporal locations of the different nodes in the combination.
Agricultural pattern analysis system
A pattern recognition system including an image gathering unit that gathers at least one digital representation of a field, an image analysis unit that pre-processes the at least one digital representation of a field, an annotation unit that provides a visualization of at least one channel for each of the at least one digital representation of the field, where the image analysis unit generates a plurality of image samples from each of the at least one digital representation of the field, and the image analysis unit splits each of the image samples into a plurality of categories.
Agricultural pattern analysis system
A pattern recognition system including an image gathering unit that gathers at least one digital representation of a field, an image analysis unit that pre-processes the at least one digital representation of a field, an annotation unit that provides a visualization of at least one channel for each of the at least one digital representation of the field, where the image analysis unit generates a plurality of image samples from each of the at least one digital representation of the field, and the image analysis unit splits each of the image samples into a plurality of categories.
Using morphological operations to process frame masks in video content
A computer implemented method can decode a frame of video data comprising an array of pixels to obtain decoded luma values and decoded chroma values corresponding to the array of pixels, and extract a frame mask based on the decoded luma values. The frame mask can include an array of mask values respectively corresponding to the array of pixels. A mask value indicates whether a corresponding pixel is in foreground or background of the frame. The method can perform a morphological operation to the frame mask to change one or more mask values to indicate their corresponding pixels are removed from the foreground and added to the background of the frame. The method can also identify foreground pixels after performing the morphological operation to the frame mask, and render a foreground image for display based on the decoded luma values and decoded chroma values of the foreground pixels.
Deep learning-based method and device for calculating overhang of battery
A deep learning-based method for calculating an overhang of a battery includes the following steps: obtaining a training sample image set; training a neural network according to the training sample image set to obtain a segmentation network model; detecting an object detection image of the battery to be detected according to the segmentation network model to obtain a corresponding first binarized image; obtaining top coordinates of each of a positive electrode and a negative electrode of the battery to be detected according to the first binarized image; and calculating the overhang of the battery to be detected according to the top coordinates.
Deep learning-based method and device for calculating overhang of battery
A deep learning-based method for calculating an overhang of a battery includes the following steps: obtaining a training sample image set; training a neural network according to the training sample image set to obtain a segmentation network model; detecting an object detection image of the battery to be detected according to the segmentation network model to obtain a corresponding first binarized image; obtaining top coordinates of each of a positive electrode and a negative electrode of the battery to be detected according to the first binarized image; and calculating the overhang of the battery to be detected according to the top coordinates.