Patent classifications
G06F18/2411
Image classification system
A method comprising: obtaining an image; identifying a rotation angle for the image by processing the image with a first neural network; rotating the image by the identified rotation angle to generate a rotated image; classifying the image with a second neural network; and outputting an indication of an outcome of the classification, wherein the first neural network is trained, at least in part, based on a categorical distance between training data and an output that is produced by the first neural network.
Data model generation using generative adversarial networks
Methods for generating data models using a generative adversarial network can begin by receiving a data model generation request by a model optimizer from an interface. The model optimizer can provision computing resources with a data model. As a further step, a synthetic dataset for training the data model can be generated using a generative network of a generative adversarial network, the generative network trained to generate output data differing at least a predetermined amount from a reference dataset according to a similarity metric. The computing resources can train the data model using the synthetic dataset. The model optimizer can evaluate performance criteria of the data model and, based on the evaluation of the performance criteria of the data model, store the data model and metadata of the data model in a model storage. The data model can then be used to process production data.
Data model generation using generative adversarial networks
Methods for generating data models using a generative adversarial network can begin by receiving a data model generation request by a model optimizer from an interface. The model optimizer can provision computing resources with a data model. As a further step, a synthetic dataset for training the data model can be generated using a generative network of a generative adversarial network, the generative network trained to generate output data differing at least a predetermined amount from a reference dataset according to a similarity metric. The computing resources can train the data model using the synthetic dataset. The model optimizer can evaluate performance criteria of the data model and, based on the evaluation of the performance criteria of the data model, store the data model and metadata of the data model in a model storage. The data model can then be used to process production data.
Disaggregation system
A computing device determines a disaggregated solution vector of a plurality of variables. A first value is computed for a known variable using a predefined density distribution function, and a second value is computed for an unknown variable using the computed first value, a predefined correlation value, and a predefined aggregate value. The predefined correlation value indicates a correlation between the known variable and the unknown variable. A predefined number of solution vectors is computed by repeating the first value and the second value computations. A solution vector is the computed first value and the computed second value. A centroid vector is computed from solution vectors computed by repeating the computations. A predefined number of closest solution vectors to the computed centroid vector are determined from the solution vectors. The determined closest solution vectors are output.
Disaggregation system
A computing device determines a disaggregated solution vector of a plurality of variables. A first value is computed for a known variable using a predefined density distribution function, and a second value is computed for an unknown variable using the computed first value, a predefined correlation value, and a predefined aggregate value. The predefined correlation value indicates a correlation between the known variable and the unknown variable. A predefined number of solution vectors is computed by repeating the first value and the second value computations. A solution vector is the computed first value and the computed second value. A centroid vector is computed from solution vectors computed by repeating the computations. A predefined number of closest solution vectors to the computed centroid vector are determined from the solution vectors. The determined closest solution vectors are output.
Processing Multimodal User Input for Assistant Systems
In one embodiment, a method includes receiving at a head-mounted device a speech input from a user and a visual input captured by cameras of the head-mounted device, wherein the visual input comprises subjects and attributes associated with the subjects, and wherein the speech input comprises a co-reference to one or more of the subjects, resolving entities corresponding to the subjects associated with the co-reference based on the attributes and the co-reference, and presenting a communication content responsive to the speech input and the visual input at the head-mounted device, wherein the communication content comprises information associated with executing results of tasks corresponding to the resolved entities.
Techniques for modeling temporal distortions when predicting perceptual video quality
In various embodiments, a prediction application computes a quality score for re-constructed visual content that is derived from visual content. The prediction application generates a frame difference matrix based on two frames included in the re-constructed video content. The prediction application then generates a first entropy matrix based on the frame difference matrix and a first scale. Subsequently, the prediction application computes a first value for a first temporal feature based on the first entropy matrix and a second entropy matrix associated with both the visual content and the first scale. The prediction application computes a quality score for the re-constructed video content based on the first value, a second value for a second temporal feature associated with a second scale, and a machine learning model that is trained using subjective quality scores. The quality score indicates a level of visual quality associated with streamed video content.
Techniques for modeling temporal distortions when predicting perceptual video quality
In various embodiments, a prediction application computes a quality score for re-constructed visual content that is derived from visual content. The prediction application generates a frame difference matrix based on two frames included in the re-constructed video content. The prediction application then generates a first entropy matrix based on the frame difference matrix and a first scale. Subsequently, the prediction application computes a first value for a first temporal feature based on the first entropy matrix and a second entropy matrix associated with both the visual content and the first scale. The prediction application computes a quality score for the re-constructed video content based on the first value, a second value for a second temporal feature associated with a second scale, and a machine learning model that is trained using subjective quality scores. The quality score indicates a level of visual quality associated with streamed video content.
ACTION RECOGNITION USING IMPLICIT POSE REPRESENTATIONS
A computer-implemented method of recognition of actions performed by individuals includes: by one or more processors, obtaining images including at least a portion of an individual by the one or more processors, based on the images, generating implicit representations of poses of the individual in the images; and by the one or more processors, determining an action performed by the individual and captured in the images by classifying the implicit representations of the poses of the individual.
ACTION RECOGNITION USING IMPLICIT POSE REPRESENTATIONS
A computer-implemented method of recognition of actions performed by individuals includes: by one or more processors, obtaining images including at least a portion of an individual by the one or more processors, based on the images, generating implicit representations of poses of the individual in the images; and by the one or more processors, determining an action performed by the individual and captured in the images by classifying the implicit representations of the poses of the individual.