Patent classifications
G06V10/768
DISPLAY CONTROL METHOD AND DEVICE
A method includes acquiring an image, acquiring display orders of a plurality of object data that respectively correspond to a plurality of reference objects based on correspondence information in which a reference object is associated with an object data that corresponds to the reference object and a display order of the object data, determining, among the plurality of object data, object data that corresponds to a display subject based on the display orders of the plurality of object data, executing a process that generates display information for displaying the object data that is the display subject, controlling a display to display the object data that is the display subject based on an execution result of the process, and performing the executing of the process for another object data, and the controlling of the display based on the another object data, the another object data being a next display subject.
CONTEXTUALLY DISAMBIGUATING QUERIES
Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for contextually disambiguating queries are disclosed. In an aspect, a method includes receiving an image being presented on a display of a computing device and a transcription of an utterance spoken by a user of the computing device, identifying a particular sub-image that is included in the image, and based on performing image recognition on the particular sub-image, determining one or more first labels that indicate a context of the particular sub-image. The method also includes, based on performing text recognition on a portion of the image other than the particular sub-image, determining one or more second labels that indicate the context of the particular sub-image, based on the transcription, the first labels, and the second labels, generating a search query, and providing, for output, the search query.
METHOD AND SYSTEM FOR CLASSIFYING IMAGES USING IMAGE EMBEDDING
There is described a computer-implemented method and system for classifying images, the computer-implemented method comprising: receiving an image to be classified, generating a vector representation of the image to be classified using an image embedding method, comparing the vector representation of the image to predefined vector representations of the predefined image categories, and identifying a relevant category amongst the predefined image categories based on the comparison, the relevant category being associated with the image to be classified and outputting the relevant category.
SYSTEMS AND METHODS FOR OBJECT DETECTION USING A GEOMETRIC SEMANTIC MAP BASED ROBOT NAVIGATION
This disclosure relates generally to systems and methods for object detection using a geometric semantic map based robot navigation using an architecture to empower a robot to navigate an indoor environment with logical decision making at each intermediate stage. The decision making is further enhanced by knowledge on actuation capability of the robots and that of scenes, objects and their relations maintained in an ontological form. The robot navigates based on a Geometric Semantic map which is a relational combination of geometric and semantic map. In comparison to traditional approaches, the robot's primary task here is not to map the environment, but to reach a target object. Thus, a goal given to the robot is to find an object in an unknown environment with no navigational map and only egocentric RGB camera perception.
APPARATUS, METHOD AND COMPUTER-READABLE STORAGE MEDIUM FOR DETECTING OBJECTS IN A VIDEO SIGNAL BASED ON VISUAL EVIDENCE USING AN OUTPUT OF A MACHINE LEARNING MODEL
Detections in video frames of a video signal, which are output from a machine learning model, are associated to generate a detection chain. Display of a detection in the video signal is caused based on a position of the detection in the detection chain, the confidence value of the detection and the location of the detection.
MOBILE SENSING SYSTEM FOR CROP MONITORING
Described herein are mobile sensing units for capturing raw data corresponding to certain characteristics of plants and their growing environment. Also described are computer devices and related methods for collecting user inputs, generating information relating to the plants and/or growing environment based on the raw data and user inputs, and displaying same.
Electronic apparatus and method for assisting with driving of vehicle
An electronic apparatus and method for assisting with driving of a vehicle are provided. The electronic apparatus includes: a processor configured to execute one or more instructions stored in a memory, to: obtain a surrounding image of the vehicle via at least one sensor, recognize an object from the obtained surrounding image, obtain three-dimensional (3D) coordinate information for the object by using the at least one sensor, determine a number of planar regions constituting the object, based on the 3D coordinate information corresponding to the object, determine whether the object is a real object, based on the number of planar regions constituting the object, and control a driving operation of the vehicle based on a result of the determining whether the object is the real object.
METHOD AND SYSTEM FOR CONFIDENCE LEVEL DETECTION FROM EYE FEATURES
State of art techniques attempt in extracting insights from eye features, specifically pupil with focus on behavioral analysis than on confidence level detection. Embodiments of the present disclosure provide a method and system for confidence level detection from eye features using ML based approach. The method enables generating overall confidence level label based on the subject's performance during an interaction, wherein the interaction that is analyzed is captured as a video sequence focusing on face of the subject. For each frame facial features comprising an Eye-Aspect ratio, a mouth movement, Horizontal displacements, Vertical displacements, Horizontal Squeezes and Vertical Peaks, are computed, wherein HDs, VDs, HSs and VPs are features that are derived from points on eyebrow with reference to nose tip of the detected face. This is repeated for all frames in the window. A Bi-LSTM model is trained using the facial features to derive confidence level of the subject.
Data model generation using generative adversarial networks
Methods for generating data models using a generative adversarial network can begin by receiving a data model generation request by a model optimizer from an interface. The model optimizer can provision computing resources with a data model. As a further step, a synthetic dataset for training the data model can be generated using a generative network of a generative adversarial network, the generative network trained to generate output data differing at least a predetermined amount from a reference dataset according to a similarity metric. The computing resources can train the data model using the synthetic dataset. The model optimizer can evaluate performance criteria of the data model and, based on the evaluation of the performance criteria of the data model, store the data model and metadata of the data model in a model storage. The data model can then be used to process production data.
CONTEXT-AIDED MACHINE VISION
Various embodiments herein each include at least one of systems, methods, software, and data structures for context-aided machine vision. For example, one method embodiment includes identifying a customer in a shopping area and maintaining an item bin in a computing system of data identifying items the customer has picked up for purchase. This method further includes receiving an image of the customer holding an item and performing item identification processing on the image to identify the item the customer is holding. The item identification processing may be performed based in part on a stored shopping history of the customer indicating items the customer is more likely to purchase. The identified item is then added to the item bin of the customer.