G06V30/19013

TEXT DETECTION, CARET TRACKING, AND ACTIVE ELEMENT DETECTION
20210342625 · 2021-11-04 · ·

Detection of typed and/or pasted text, caret tracking, and active element detection for a computing system are disclosed. The location on the screen associated with a computing system where the user has been typing or pasting text, potentially including hot keys or other keys that do not cause visible characters to appear, can be identified and the physical position on the screen where typing or pasting occurred can be provided based on the current resolution of where one or more characters appeared, where the cursor was blinking, or both. This can be done by identifying locations on the screen where changes occurred and performing text recognition and/or caret detection on these locations. The physical position of the typing or pasting activity allows determination of an active or focused element in an application displayed on the screen.

Information processing apparatus and non-transitory computer readable medium

An information processing apparatus includes a processor configured to: acquire, from results of character recognition performed on a target image including character strings, attribute information indicating an attribute to which a key character string and a value character string belong, the key character string as a character string specified beforehand as a key and the value character string as a character string indicating a value corresponding to the key character string; acquire by using the attribute information the key character string corresponding to the value character string extracted from the results of the character recognition; and output the key character string and the value character string corresponding to the key character string.

RECOGNITION AND INDICATION OF DISCRETE PATTERNS WITHIN A SCENE OR IMAGE

Selection of on optical pattern in a scene is identified by overlaying, on a display, an indicator of a detected optical pattern identifying a location of the optical pattern in one or more images, receiving a user input on the display at a position that does not overlap the location of the optical pattern, and presenting information related to the optical pattern, based on receiving the user input, even though the position of user input did not overlap the location of the optical pattern. The user input can be received at a detached selection indicator and/or using an adaptive input area.

Text detection, caret tracking, and active element detection
11630549 · 2023-04-18 · ·

Detection of typed and/or pasted text, caret tracking, and active element detection for a computing system are disclosed. The location on the screen associated with a computing system where the user has been typing or pasting text, potentially including hot keys or other keys that do not cause visible characters to appear, can be identified and the physical position on the screen where typing or pasting occurred can be provided based on the current resolution of where one or more characters appeared, where the cursor was blinking, or both. This can be done by identifying locations on the screen where changes occurred and performing text recognition and/or caret detection on these locations. The physical position of the typing or pasting activity allows determination of an active or focused element in an application displayed on the screen.

MULTIPLE INPUT MACHINE LEARNING FRAMEWORK FOR ANOMALY DETECTION

A method that includes extracting image features of a document image, executing an optical character recognition (OCR) engine on the document image to obtain OCR output, and extracting OCR features from the OCR output. The method further includes executing an anomaly detection model using features including the OCR features and the image features to generate anomaly score, and presenting anomaly score.

Mobile supplementation, extraction, and analysis of health records

A system, method, and mobile device application are configured to capture, with a mobile device, a document such as a next generation sequencing (NGS) report that includes NGS medical information about a genetically sequenced patient. The method includes receiving, from a mobile device, an image of a medical document comprising NGS medical information of the patient, extracting a first region from the image, extracting NGS medical information of the patient from the first region into a structured dataset, the extracted NGS medical information including at least one RNA expression, correlating a portion of the extracted NGS medical information that includes the at least one RNA expression with summarized medical information from a cohort of patients similar to the patient, and generating, for display on the mobile device, a clinical decision support report comprising the summarized medical information.

SYSTEMS AND METHODS FOR LOW COMPUTE DEPTH MAP GENERATION

Systems and methods are provided performing for low compute depth map generation by implementing acts of obtaining a stereo pair of images of a scene, downsampling the stereo pair of images, generating a depth map by stereo matching the downsampled stereo pair of images, and generating an upsampled depth map based on the depth map using an edge-preserving filter for obtaining at least some data of at least one image of the stereo pair of images.

DETECTION OF MANUAL ENTRY ERROR

Systems and computer-implemented methods disclosed herein relate to detecting errors in manually entered data. In one embodiment, the system can identify a named entity automatically from a conversation between a customer and service agent with a named entity recognition model that employs natural language processing and machine learning to detect a word or string of words in the conversation that corresponds to a named entity category. In another embodiment, the system can determine whether data entered into a field on a service platform by the service agent includes an error by comparing the data entered with the named entity. In another embodiment, the system can transmit an alert to the service agent through the service platform when there is a mismatch between the named entity and the data entered.

METHOD, DEVICE, AND SYSTEM FOR OUTPUTTING DESCRIPTION OF PATENT REFERENCE SIGN

A method for outputting a drawing reference number description regarding a patent drawing reference number according to an embodiment of the present disclosure may include recognizing a size of a patent drawing and a position of a drawing reference number included in the patent drawing and acquiring a relative position coordinate of the drawing reference number in the patent drawing; setting a relative position coordinate of the drawing reference number description corresponding to the drawing reference number based on the acquired relative position coordinate; and outputting the drawing reference number description on the set relative position coordinate, thereby outputting such that the drawing reference number description corresponds to the drawing reference number.

DRUG IDENTIFICATION DEVICE, DRUG IDENTIFICATION METHOD AND PROGRAM, DRUG IDENTIFICATION SYSTEM, DRUG LOADING TABLE, ILLUMINATION DEVICE, IMAGING ASSISTANCE DEVICE, TRAINED MODEL, AND LEARNING DEVICE

A region of a drug to be identified is detected from a captured image generated by imaging the drug to be identified that is imparted with engraved mark and/or print. The region of the drug to be identified in the captured image is processed to acquire an engraved mark and print extraction image that is an extracted image of the engraved mark and/or print of the drug to be identified. The engraved mark and print extraction image is input, and a drug type of the drug to be identified is inferred to acquire a candidate of the drug type of the drug to be identified.