Patent classifications
G06F18/25
DEVICE FOR DETERMINING A FACE OF A DICE RESTING ON A SURFACE ALLOWING AN OPTICAL SIGNAL TO PASS
A device for determining a face of a dice resting on a surface allowing an optical signal to pass, the dice being composed of a plurality of faces each including a visual marking uniquely identifying the face, and the device including: illumination means for illuminating a face of a dice through the surface, the illumination means being placed under the surface and oriented in the direction of the surface, the illumination means including a plurality of optical signal sources disposed at various positions under the surface; means for acquiring at least one image of the optical signals reflected by the face of the dice resting on the surface, the acquisition means being placed under the surface and being facing the surface; and an analysis unit including means for processing the image to determine the face of the dice resting on the surface allowing the optical signal to pass.
Facial beauty prediction method and device based on multi-task migration
Disclosed are a facial beauty prediction method and device based on multi-task migration. The method includes: performing similarity measurement based on a graph structure on a plurality of tasks to obtain an optimal combination of the plurality of tasks; constructing a facial beauty prediction model including a feature sharing layer based on the optimal combination; migrating feature parameters of an existing large-scale facial image network to the feature sharing layer of the facial beauty prediction model; inputting facial images for training to pre-train the facial beauty prediction model; and inputting a facial image to be tested to the trained facial beauty prediction model to obtain facial recognition results.
ENVIRONMENTAL MODEL BASED ON AUDIO
A method for providing an audio-based model of an environment of a vehicle, the method may include obtaining, during a driving session of a vehicle, sensed information about the environment of the vehicle; wherein the sensed information may include sensed audio information. The sensed information may also include at least one type of non-audio sensed information; and generating an audio-based model of the environment based, at least in part, on the sensed audio information.
MEMORY-AUGMENTED GRAPH CONVOLUTIONAL NEURAL NETWORKS
System and method for processing a graph that defines a set of nodes and a set of edges, the nodes each having an associated set of node attributes, the edges each representing a relationship that connects two respective nodes, comprising: generating a first node embedding for each node by: generating, for the node and each of a plurality of neighbour nodes, a respective first edge attribute defining a respective relationship type between the node and the neighbour node based on the node attributes of the node and the node attributes of the neighbour node; generating a first neighborhood vector that aggregates information from the generated first edge attributes and the node attributes of the neighbour nodes; generating the first node embedding based on the node attributes of the node and the generated first neighborhood vector.
FEATURE ENGINEERING USING INTERACTIVE LEARNING BETWEEN STRUCTURED AND UNSTRUCTURED DATA
A concept associated with a feature used in machine learning model can be determined, the feature extracted from a first data source. A second data source containing the concept can be identified. An additional feature can be generated by performing a natural language processing on the second data source. The feature and the additional feature can be merged. A second machine learning model can be generated, which use the merged feature. A prediction result of the first machine learning model can be compared with a prediction result of the second machine learning model relative to ground truth data, to evaluate effective of the merged feature. Based on the evaluated effectiveness, the feature can be augmented with the merged feature in machine learning.
MODEL OPTIMIZATION METHOD, ELECTRONIC DEVICE, AND COMPUTER PROGRAM PRODUCT
Embodiments of the present disclosure relate to a model optimization method, an electronic device, and a computer program product. This method includes: determining an initial learning rate combination for a deep learning model, wherein the initial learning rate combination includes a plurality of learning rates, each learning rate being determined for one of a plurality of layers of the deep learning model, and the plurality of learning rates including static learning rates and dynamic learning rates; and adjusting the initial learning rate combination to obtain a target learning rate combination, wherein an accuracy rate achieved when the target learning rate combination is used to train the deep learning model is higher than or equal to a first threshold accuracy rate. With the technical solution of the present disclosure, a deep learning model can be optimized by setting learning rates for each layer of the deep learning model.
DETECTION OF PLANT DISEASES WITH MULTI-STAGE, MULTI-SCALE DEEP LEARNING
A computer system is provided comprising a classification model management server computer configured, by instructions, to: receive a new image from a user device; apply a first digital model to first regions within the new image for classifying each of the first regions into a particular class; apply a second digital model to second regions within the new image for classifying each of the second regions into a particular class; and transmit classification data related to the class of the first regions and the class of the second regions to the user device. In connection therewith, the second regions each generally correspond to a combination of multiple first regions.
Interactive assistant
An interactive troubleshooting assistant and method for troubleshooting a system in real time to repair (fix) one or more problems in a system is disclosed. The interactive troubleshooting assistant and method may include receiving multimodal inputs from sensors, wearable devices, a person, etc. that may be input into a feature extractor including attention layers and pre-processing units of a cloud computing system hosted by one or more servers, such as a private cloud system. A pre-processing unit converts the raw multimodal input into a structed form so that an attention layer can give weights to features provided by the pre-processing unit according to their importance. The weighted extracted features may be provided to an actions predictor. The actions predictor generates the most suitable action based on the weighted extracted features generated by the feature extractor based on the multimodal inputs. After the most suitable action is performed, the interactive troubleshooting assistant considers new information from multimodal inputs so that the interactive troubleshooting assistant can provide the next recommended action. The interactive troubleshooting assistant may repeat these operations until the repair is completed.
AUGMENTING AUDIENCE MEMBER EMOTES IN LARGE-SCALE ELECTRONIC PRESENTATION
A presentation service generates an audience interface for an electronic presentation. The audience interface may simulate an in-person presentation, including features such as a central presenter and seat locations for audience members. The audience members may select emotes which may be displayed in the audience interface. The emotes may indicate the audience members' opinion of the content being presented. The presentation service may enable chats between multiple audience members, grouping of audience members private rooms, and other virtual simulations of functions corresponding to in-person presentations.
Method, apparatus, terminal, and storage medium for training model
This application disclose a method for training a model performed at a computing device. The method includes: acquiring a template image and a test image; invoking a first object recognition model to process a feature of a tracked object in the template image to obtain a first reference response, and a second object recognition model to process the feature in the template image to obtain a second reference response; invoking the first model to process a feature of a tracked object in the test image to obtain a first test response, and the second model to process the feature to obtain a second test response; tracking the first test response to obtain a tracking response of the tracked object; and updating the first object recognition model based on differences between the first and second reference responses, that between the first and second test responses, and that between a tracking label and the tracking response.