Patent classifications
G06V10/464
IMAGE PROCESSING DEVICE, IMAGE PROCESSING METHOD, AND PROGRAM
An image processing device includes a search scenario acquisition unit and a search unit. The search scenario acquisition unit acquires a search scenario in which one or more search processes of searching for an image using an image feature based on a target image, which is an image of a search target, in a search condition are combined. The search unit searches for a matching image, which is an image to be searched, using a combination of the search processes represented by the search scenario and outputs search result information representing an area in the matching image detected through the search.
METHOD AND SYSTEM FOR CHECKING DATA GATHERING CONDITIONS ASSOCIATED WITH IMAGE-DATA DURING AI ENABLED VISUAL-INSPECTION PROCESS
A method and system for checking data gathering conditions or image capturing conditions associated with images during AI based visual-inspection process. The method comprises generating a first representative (FR1) image for a first group of images and a second representative image (FR2) for a second group of images. A difference image data is generated between FR1 image and the FR2 image based on calculating difference between luminance values of pixels with same coordinate values. Thereafter, one or more of a plurality of white pixels or intensity-values are determined within the difference image based on acquiring difference image data formed of luminance difference-values of pixels. An index representing difference of data-capturing conditions across the FR1 image and the FR2 image is determined, said index having been determined at least based on the plurality of white pixels or intensity-values, for example, based on application of a plurality of AI or ML techniques.
Advanced response processing in web data collection
ADVANCED RESPONSE PROCESSING IN WEB DATA COLLECTION discloses processor-implemented apparatuses, methods, and systems of processing unstructured raw HTML responses collected in the context of a data collection service, the method comprising, in one embodiment, receiving raw unstructured HTML documents and extracting text data with associated meta information that may comprise style and formatting information. In some embodiments data field tags and values may be assigned to the text blocks extracted, classifying the data based on the processing of Machine Learning algorithms. Additionally, blocks of extracted data may be grouped and re-grouped together and presented as a single data point. In another embodiment the system may aggregate and present the text data with the associated meta information in a structured format. In certain embodiments the Machine Learning model may be a model trained on a pre-created training data set labeled manually or in an automatic fashion.
Methods and Systems for Generating Composite Image Descriptors
An illustrative image descriptor generation system generates a descriptor listing that includes a plurality of image descriptors corresponding to different feature points included within an image. Based on the descriptor listing, the system generates a geometric map representing the plurality of image descriptors in accordance with respective geometric positions of the corresponding feature points of the image descriptors within the image. Based on the geometric map, the system determines a proximity listing for a primary image descriptor within the plurality of image descriptors. The proximity listing indicates a subset of image descriptors that are geometrically proximate to the primary image descriptor within the image. Based on the proximity listing, the system selects a secondary image descriptor from the subset of image descriptors and combines the primary and secondary image descriptors to form a composite image descriptor. Corresponding methods and systems are also disclosed.
Systems and Methods for Collaborative Edge Computing
An edge computing system configured to dynamically offload tasks from a user device to an edge device. The edge device may receive a request to download or run a software application suitable processing a sensory feed collected by the user device. The edge device may determine operating requirements of the software application, determine the internal capabilities of the edge device, and compare the determined operating requirements of the software application to the determined capabilities of the edge device to determine whether the edge device is capable of running the entire software application. The edge device may also determine whether the edge device is capable of running a portion of the software application in response to determining that the edge device is not capable of running the entire software application. The edge device may send a message identifying portions that could be run on the edge device.
Systems and Methods for Collaborative Edge Computing
An edge computing system configured to dynamically offload tasks from a user device to an edge device. The edge computing system may receive a processed sensory feed from the user device, analyze the received processed sensory feed, and generate mapper output results. The edge computing system may compare the generated mapper output results to information received from a datastore, identify a correlation between a feature included in the received processed sensory feed and a feature included in the received information, and determine whether a confidence value associated with the identified correlation exceeds a threshold value. The edge computing system may further process the received processed sensory feed locally in the edge computing system or send the received processed sensory feed to a cloud component for further processing based on whether the confidence value exceeds the threshold value.
Adaptive Pattern Recognition for a Sensor Network
Embodiments match sensor data output by a sensor to a trained pattern. Embodiments form a plurality of windows of an identified pattern from the sensor data, each of the plurality of windows having a substantially equal window length to a length of the trained pattern. For each of the windows, embodiments generate a corresponding first Symbolic Aggregate approximation (“SAX”) word, determine a Hamming distance between the first SAX word and a second SAX word corresponding to the trained pattern, and determine a final distance score based on coefficients between the first SAX word and the second SAX word. For each of the windows, embodiments determine a number of positions in the first SAX word that do not contribute to the final distance score, update the Hamming distance after eliminating the number of positions and determine an average distance based on the final distance score and the updated Hamming distance.
Methods and systems for generating composite image descriptors
An illustrative image descriptor generation system generates a descriptor listing that includes a plurality of image descriptors corresponding to different feature points included within an image. Based on the descriptor listing, the system generates a geometric map representing the plurality of image descriptors in accordance with respective geometric positions of the corresponding feature points of the image descriptors within the image. Based on the geometric map, the system determines a proximity listing for a primary image descriptor within the plurality of image descriptors. The proximity listing indicates a subset of image descriptors that are geometrically proximate to the primary image descriptor within the image. Based on the proximity listing, the system selects a secondary image descriptor from the subset of image descriptors and combines the primary and secondary image descriptors to form a composite image descriptor. Corresponding methods and systems are also disclosed.
PIXEL CORRESPONDENCE VIA PATCH-BASED NEIGHBORHOOD CONSENSUS
One example provides a computing system comprising a storage machine storing instructions executable by a logic machine to extract features from a source and target images to form source and target feature maps, form a correlation map comprising a plurality of similarity scores, form an initial correspondence map comprising initial mappings between pixels of the source feature map and corresponding pixels of the target feature map, refine the initial correspondence map by, for each of one or more pixels of the source feature map, for each of a plurality of candidate correspondences, inputting a four-dimensional patch into a trained scoring function, the trained scoring function being configured to output a correctness score, and selecting a refined correspondence based at least upon the correctness scores, and output a refined correspondence map comprising a refined correspondence for each of the one or more pixels of the source feature map.
Systems, methods, and apparatuses for implementing a self-supervised chest x-ray image analysis machine-learning model utilizing transferable visual words
Not only is annotating medical images tedious and time consuming, but it also demands costly, specialty-oriented expertise, which is not easily accessible. To address this challenge, a new self-supervised framework is introduced: TransVW (transferable visual words), exploiting the prowess of transfer learning with convolutional neural networks and the unsupervised nature of visual word extraction with bags of visual words, resulting in an annotation-efficient solution to medical image analysis. TransVW was evaluated using NIH ChestX-ray14 to demonstrate its annotation efficiency. When compared with training from scratch and ImageNet-based transfer learning, TransVW reduces the annotation efforts by 75% and 12%, respectively, in addition to significantly accelerating the convergence speed. More importantly, TransVW sets new records: achieving the best average AUC on all 14 diseases, the best individual AUC scores on 10 diseases, and the second best individual AUC scores on 3 diseases. This performance is unprecedented, because heretofore no self-supervised learning method has outperformed ImageNet-based transfer learning and no annotation reduction has been reported for self-supervised learning. These achievements are contributable to a simple yet powerful observation: The complex and recurring anatomical structures in medical images are natural visual words, which can be automatically extracted, serving as strong yet free supervision signals for CNNs to learn generalizable and transferable image representation via self-supervision.