Patent classifications
G06V30/19173
TRANSPARENCY DETECTION METHOD BASED ON MACHINE VISION
Disclosed is a transparency detecting method based on machine vision. The transparency detecting method includes 1) operating a Secchi disk to start the water transparency measurement, and turning on the camera for shooting; 2). determining a critical position of the Secchi disk; 3) identifying a water ruler and calculating a reading of the water ruler; 4) outputting and displaying the calculated reading.
AUTOMATED TELLER MACHINE FOR DETECTING SECURITY VULNERABILITIES BASED ON DOCUMENT NOISE REMOVAL
An Automated Teller Machine (ATM) for detecting security vulnerabilities by removing noise artifacts from documents receives a transaction request when a document is inserted into the ATM, where the document contains a noise artifact at least partially obstructing a portion of the document. The ATM generates an image of the document, where the image displays at least one data item comprising a sender's name, a receiver's name, and a number representing an amount. The ATM determines whether the noise artifact obstructs at least partially one data item. In response to determining that the noise artifact obstructs at least partially one data item, the ATM generates a test clean image of the document by removing the noise artifact from the image. In response to determining that the noise artifact is removed, the ATM approves the transaction request.
Embedding human labeler influences in machine learning interfaces in computing environments
A mechanism is described for facilitating embedding of human labeler influences in machine learning interfaces in computing environments, according to one embodiment. A method of embodiments, as described herein, includes detecting sensor data via one or more sensors of a computing device, and accessing human labeler data at one or more databases coupled to the computing device. The method may further include evaluating relevance between the sensor data and the human labeler data, where the relevance identifies meaning of the sensor data based on human behavior corresponding to the human labeler data, and associating, based on the relevance, human labeler data with the sensor data to classify the sensor data as labeled data. The method may further include training, based on the labeled data, a machine learning model to extract human influences from the labeled data, and embed one or more of the human influences in one or more environments representing one or more physical scenarios involving one or more humans.
METHOD FOR TRAINING TEXT CLASSIFICATION MODEL, ELECTRONIC DEVICE AND STORAGE MEDIUM
A method for training a text classification model and an electronic device are provided. The method may include: acquiring a set of to-be-trained images, the set of to-be-trained images including at least one sample image; determining predicted position information and predicted attribute information of each text line in each sample image based on each sample image; and training to obtain the text classification model, based on the annotation position information and the annotation attribute information of each text line in each sample image, and the predicted position information and the predicted attribute information of each text line in each sample image, and the text classification model is used to detect attribute information of each text line in an to-be-recognized image.
Calculation practicing method, system, electronic device and computer readable storage medium
The disclosure provides a calculation practicing method, a system, an electronic device and a computer readable storage medium, the calculation practicing method includes: providing a calculation question; identifying the type and content of the calculation question; generating an answer area according to the type and content of the calculation question; receiving an answering operation in which the user inputs the answer string for the calculation question in the answer area; identifying the answer string inputted by the user; and determining whether each of the answer characters in the answer string is correct, if there is an incorrect answer character, it will be marked, so that the calculation practice can be realized through the electronic device, which is convenient for students to carry out training.
Utilizing machine learning models, position based extraction, and automated data labeling to process image-based documents
A device may receive image data that includes an image of a document and lexicon data identifying a lexicon, and may perform an extraction technique on the image data to identify at least one field in the document. The device may utilize form segmentation to automatically generate label data identifying labels for the image data, and may process the image data, the label data, and data identifying the at least one field, with a first model, to identify visual features. The device may process the image data and the visual features, with a second model, to identify sequences of characters, and may process the image data and the sequences of characters, with a third model, to identify strings of characters. The device may compare the lexicon data and the strings of characters to generate verified strings of characters that may be utilized to generate a digitized document.
APPARATUS AND METHOD FOR GENERATING A SCHEMA
An apparatus and method for generating a schema, the apparatus comprising at least a processor and a memory communicatively connected to the at least a processor, the memory containing instructions configuring the at least a processor to display, at a graphical control interface, a content field window, receive, as a function of the content field window, a criterion element, and generate a schema as a function of the criterion element.
Using few shot learning on recognition system for character image in industrial processes
An artificial intelligence optical character image recognition system and method, using few shot learning on recognition system for character image in industrial processes, mainly including: preparing two or more identical neural network architecture units, inputting similar or different character images respectively, and comparing the calculation results to see if the weights are similar. If the similarity reaches the set standard value, they are classified as the same type of character, otherwise different. Through such procedures, training samples in the storage unit are gradually divided into settings of character sets with different contextual meanings, becoming a complete AI OCR system. It can increase training sample data by comparing characters, without increasing the training set. Simultaneously, it can improve the flexibility of recognizing test characters.
Teaching GAN (generative adversarial networks) to generate per-pixel annotation
A method and apparatus for joint image and per-pixel annotation synthesis with a generative adversarial network (GAN) are provided. The method includes: by inputting data to a generative adversarial network (GAN), obtaining a first image from the GAN; inputting, to a decoder, a first feature value that is obtained from at least one intermediate layer of the GAN according to the inputting of the data to the GAN; and obtaining a first semantic segmentation mask from the decoder according to the inputting of the first feature value to the decoder.
Learning user interface controls via incremental data synthesis
A User Interface (UI) interface object detection system employs an initial dataset comprising a set of images, that may include synthesized images, to train a Machine Learning (ML) engine to generate an initial trained model. A data point generator is employed to generate an updated synthesized image set which is used to further train the ML engine. The data point generator may employ images generated by an application program as a reference by which to generate the updated synthesized image set. The images generated by the application program may be tagged in advance. Alternatively, or in addition, the images generated by the application program may be captured dynamically by a user using the application program.